var/home/core/zuul-output/0000755000175000017500000000000015111627671014534 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015111634160015467 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004340125515111634151017677 0ustar rootrootNov 26 16:52:33 crc systemd[1]: Starting Kubernetes Kubelet... Nov 26 16:52:33 crc restorecon[4679]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:33 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Nov 26 16:52:34 crc restorecon[4679]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Nov 26 16:52:34 crc kubenswrapper[4956]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 16:52:34 crc kubenswrapper[4956]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Nov 26 16:52:34 crc kubenswrapper[4956]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 16:52:34 crc kubenswrapper[4956]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 16:52:34 crc kubenswrapper[4956]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Nov 26 16:52:34 crc kubenswrapper[4956]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.735000 4956 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740279 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740313 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740323 4956 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740333 4956 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740342 4956 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740350 4956 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740359 4956 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740368 4956 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740379 4956 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740389 4956 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740397 4956 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740405 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740413 4956 feature_gate.go:330] unrecognized feature gate: Example Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740421 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740429 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740437 4956 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740445 4956 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740456 4956 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740467 4956 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740475 4956 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740483 4956 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740492 4956 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740500 4956 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740508 4956 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740516 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740524 4956 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740532 4956 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740542 4956 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740550 4956 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740558 4956 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740567 4956 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740576 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740584 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740594 4956 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740604 4956 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740613 4956 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740622 4956 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740631 4956 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740639 4956 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740648 4956 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740656 4956 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740665 4956 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740673 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740680 4956 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740690 4956 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740698 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740706 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740713 4956 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740721 4956 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740730 4956 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740738 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740747 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740756 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740764 4956 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740771 4956 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740779 4956 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740787 4956 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740795 4956 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740802 4956 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740809 4956 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740817 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740825 4956 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740832 4956 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740840 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740848 4956 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740855 4956 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740886 4956 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740894 4956 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740906 4956 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740916 4956 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.740925 4956 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741067 4956 flags.go:64] FLAG: --address="0.0.0.0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741083 4956 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741098 4956 flags.go:64] FLAG: --anonymous-auth="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741110 4956 flags.go:64] FLAG: --application-metrics-count-limit="100" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741122 4956 flags.go:64] FLAG: --authentication-token-webhook="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741133 4956 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741145 4956 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741155 4956 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741165 4956 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741174 4956 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741184 4956 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741193 4956 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741202 4956 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741211 4956 flags.go:64] FLAG: --cgroup-root="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741220 4956 flags.go:64] FLAG: --cgroups-per-qos="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741228 4956 flags.go:64] FLAG: --client-ca-file="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741237 4956 flags.go:64] FLAG: --cloud-config="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741246 4956 flags.go:64] FLAG: --cloud-provider="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741255 4956 flags.go:64] FLAG: --cluster-dns="[]" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741267 4956 flags.go:64] FLAG: --cluster-domain="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741276 4956 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741285 4956 flags.go:64] FLAG: --config-dir="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741294 4956 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741303 4956 flags.go:64] FLAG: --container-log-max-files="5" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741315 4956 flags.go:64] FLAG: --container-log-max-size="10Mi" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741324 4956 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741334 4956 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741344 4956 flags.go:64] FLAG: --containerd-namespace="k8s.io" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741353 4956 flags.go:64] FLAG: --contention-profiling="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741363 4956 flags.go:64] FLAG: --cpu-cfs-quota="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741373 4956 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741382 4956 flags.go:64] FLAG: --cpu-manager-policy="none" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741391 4956 flags.go:64] FLAG: --cpu-manager-policy-options="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741402 4956 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741412 4956 flags.go:64] FLAG: --enable-controller-attach-detach="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741421 4956 flags.go:64] FLAG: --enable-debugging-handlers="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741430 4956 flags.go:64] FLAG: --enable-load-reader="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741439 4956 flags.go:64] FLAG: --enable-server="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741448 4956 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741459 4956 flags.go:64] FLAG: --event-burst="100" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741468 4956 flags.go:64] FLAG: --event-qps="50" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741478 4956 flags.go:64] FLAG: --event-storage-age-limit="default=0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741488 4956 flags.go:64] FLAG: --event-storage-event-limit="default=0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741497 4956 flags.go:64] FLAG: --eviction-hard="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741508 4956 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741517 4956 flags.go:64] FLAG: --eviction-minimum-reclaim="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741527 4956 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741537 4956 flags.go:64] FLAG: --eviction-soft="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741546 4956 flags.go:64] FLAG: --eviction-soft-grace-period="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741555 4956 flags.go:64] FLAG: --exit-on-lock-contention="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741564 4956 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741573 4956 flags.go:64] FLAG: --experimental-mounter-path="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741582 4956 flags.go:64] FLAG: --fail-cgroupv1="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741591 4956 flags.go:64] FLAG: --fail-swap-on="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741600 4956 flags.go:64] FLAG: --feature-gates="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741611 4956 flags.go:64] FLAG: --file-check-frequency="20s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741620 4956 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741629 4956 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741639 4956 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741647 4956 flags.go:64] FLAG: --healthz-port="10248" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741656 4956 flags.go:64] FLAG: --help="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741665 4956 flags.go:64] FLAG: --hostname-override="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741674 4956 flags.go:64] FLAG: --housekeeping-interval="10s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741683 4956 flags.go:64] FLAG: --http-check-frequency="20s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741692 4956 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741700 4956 flags.go:64] FLAG: --image-credential-provider-config="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741711 4956 flags.go:64] FLAG: --image-gc-high-threshold="85" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741720 4956 flags.go:64] FLAG: --image-gc-low-threshold="80" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741729 4956 flags.go:64] FLAG: --image-service-endpoint="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741739 4956 flags.go:64] FLAG: --kernel-memcg-notification="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741747 4956 flags.go:64] FLAG: --kube-api-burst="100" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741756 4956 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741765 4956 flags.go:64] FLAG: --kube-api-qps="50" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741775 4956 flags.go:64] FLAG: --kube-reserved="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741785 4956 flags.go:64] FLAG: --kube-reserved-cgroup="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741793 4956 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741802 4956 flags.go:64] FLAG: --kubelet-cgroups="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741811 4956 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741820 4956 flags.go:64] FLAG: --lock-file="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741829 4956 flags.go:64] FLAG: --log-cadvisor-usage="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741838 4956 flags.go:64] FLAG: --log-flush-frequency="5s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741846 4956 flags.go:64] FLAG: --log-json-info-buffer-size="0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741860 4956 flags.go:64] FLAG: --log-json-split-stream="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741892 4956 flags.go:64] FLAG: --log-text-info-buffer-size="0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741902 4956 flags.go:64] FLAG: --log-text-split-stream="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741910 4956 flags.go:64] FLAG: --logging-format="text" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741919 4956 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741928 4956 flags.go:64] FLAG: --make-iptables-util-chains="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741937 4956 flags.go:64] FLAG: --manifest-url="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741946 4956 flags.go:64] FLAG: --manifest-url-header="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741958 4956 flags.go:64] FLAG: --max-housekeeping-interval="15s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741967 4956 flags.go:64] FLAG: --max-open-files="1000000" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741978 4956 flags.go:64] FLAG: --max-pods="110" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741987 4956 flags.go:64] FLAG: --maximum-dead-containers="-1" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.741996 4956 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742007 4956 flags.go:64] FLAG: --memory-manager-policy="None" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742017 4956 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742026 4956 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742036 4956 flags.go:64] FLAG: --node-ip="192.168.126.11" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742045 4956 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742064 4956 flags.go:64] FLAG: --node-status-max-images="50" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742074 4956 flags.go:64] FLAG: --node-status-update-frequency="10s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742084 4956 flags.go:64] FLAG: --oom-score-adj="-999" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742093 4956 flags.go:64] FLAG: --pod-cidr="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742101 4956 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742115 4956 flags.go:64] FLAG: --pod-manifest-path="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742124 4956 flags.go:64] FLAG: --pod-max-pids="-1" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742133 4956 flags.go:64] FLAG: --pods-per-core="0" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742141 4956 flags.go:64] FLAG: --port="10250" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742152 4956 flags.go:64] FLAG: --protect-kernel-defaults="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742161 4956 flags.go:64] FLAG: --provider-id="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742169 4956 flags.go:64] FLAG: --qos-reserved="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742178 4956 flags.go:64] FLAG: --read-only-port="10255" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742187 4956 flags.go:64] FLAG: --register-node="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742196 4956 flags.go:64] FLAG: --register-schedulable="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742206 4956 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742226 4956 flags.go:64] FLAG: --registry-burst="10" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742235 4956 flags.go:64] FLAG: --registry-qps="5" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742244 4956 flags.go:64] FLAG: --reserved-cpus="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742257 4956 flags.go:64] FLAG: --reserved-memory="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742269 4956 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742278 4956 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742288 4956 flags.go:64] FLAG: --rotate-certificates="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742297 4956 flags.go:64] FLAG: --rotate-server-certificates="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742306 4956 flags.go:64] FLAG: --runonce="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742315 4956 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742324 4956 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742333 4956 flags.go:64] FLAG: --seccomp-default="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742342 4956 flags.go:64] FLAG: --serialize-image-pulls="true" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742351 4956 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742360 4956 flags.go:64] FLAG: --storage-driver-db="cadvisor" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742370 4956 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742379 4956 flags.go:64] FLAG: --storage-driver-password="root" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742388 4956 flags.go:64] FLAG: --storage-driver-secure="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742397 4956 flags.go:64] FLAG: --storage-driver-table="stats" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742406 4956 flags.go:64] FLAG: --storage-driver-user="root" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742415 4956 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742425 4956 flags.go:64] FLAG: --sync-frequency="1m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742434 4956 flags.go:64] FLAG: --system-cgroups="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742442 4956 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742456 4956 flags.go:64] FLAG: --system-reserved-cgroup="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742465 4956 flags.go:64] FLAG: --tls-cert-file="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742474 4956 flags.go:64] FLAG: --tls-cipher-suites="[]" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742492 4956 flags.go:64] FLAG: --tls-min-version="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742501 4956 flags.go:64] FLAG: --tls-private-key-file="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742511 4956 flags.go:64] FLAG: --topology-manager-policy="none" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742520 4956 flags.go:64] FLAG: --topology-manager-policy-options="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742529 4956 flags.go:64] FLAG: --topology-manager-scope="container" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742538 4956 flags.go:64] FLAG: --v="2" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742550 4956 flags.go:64] FLAG: --version="false" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742561 4956 flags.go:64] FLAG: --vmodule="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742574 4956 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.742584 4956 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742794 4956 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742805 4956 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742814 4956 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742858 4956 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742893 4956 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742904 4956 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742912 4956 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742922 4956 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742930 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742939 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742948 4956 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742956 4956 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742965 4956 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742973 4956 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742984 4956 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742992 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.742999 4956 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743007 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743014 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743022 4956 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743030 4956 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743038 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743046 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743054 4956 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743061 4956 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743069 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743080 4956 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743089 4956 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743099 4956 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743108 4956 feature_gate.go:330] unrecognized feature gate: Example Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743119 4956 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743128 4956 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743135 4956 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743143 4956 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743151 4956 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743159 4956 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743169 4956 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743178 4956 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743186 4956 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743194 4956 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743204 4956 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743213 4956 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743222 4956 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743230 4956 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743238 4956 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743246 4956 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743254 4956 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743261 4956 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743270 4956 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743279 4956 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743288 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743297 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743305 4956 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743314 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743322 4956 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743329 4956 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743337 4956 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743345 4956 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743353 4956 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743360 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743368 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743376 4956 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743384 4956 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743393 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743402 4956 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743410 4956 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743418 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743426 4956 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743434 4956 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743442 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.743449 4956 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.743462 4956 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.757173 4956 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.757248 4956 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757413 4956 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757446 4956 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757461 4956 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757475 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757487 4956 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757498 4956 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757509 4956 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757519 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757532 4956 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757545 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757557 4956 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757568 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757580 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757590 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757600 4956 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757610 4956 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757618 4956 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757626 4956 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757634 4956 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757641 4956 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757649 4956 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757658 4956 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757665 4956 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757673 4956 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757681 4956 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757691 4956 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757707 4956 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757727 4956 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757740 4956 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757753 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757765 4956 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757777 4956 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757788 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757799 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757815 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757827 4956 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757838 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757848 4956 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757860 4956 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757902 4956 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757913 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757925 4956 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757936 4956 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757949 4956 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757959 4956 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757970 4956 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757980 4956 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.757991 4956 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758001 4956 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758011 4956 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758021 4956 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758031 4956 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758042 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758053 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758064 4956 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758076 4956 feature_gate.go:330] unrecognized feature gate: Example Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758088 4956 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758100 4956 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758112 4956 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758124 4956 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758136 4956 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758147 4956 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758158 4956 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758172 4956 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758184 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758195 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758205 4956 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758215 4956 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758226 4956 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758236 4956 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758252 4956 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.758272 4956 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758577 4956 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758607 4956 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758624 4956 feature_gate.go:330] unrecognized feature gate: PlatformOperators Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758635 4956 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758647 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758658 4956 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758669 4956 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758680 4956 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758691 4956 feature_gate.go:330] unrecognized feature gate: GatewayAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758719 4956 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758730 4956 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758741 4956 feature_gate.go:330] unrecognized feature gate: Example Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758752 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758762 4956 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758772 4956 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758782 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758793 4956 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758803 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758814 4956 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758824 4956 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758834 4956 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758846 4956 feature_gate.go:330] unrecognized feature gate: NewOLM Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758856 4956 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758901 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758913 4956 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758923 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758934 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758945 4956 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758956 4956 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758965 4956 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758975 4956 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758985 4956 feature_gate.go:330] unrecognized feature gate: PinnedImages Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.758995 4956 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759006 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759019 4956 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759030 4956 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759040 4956 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759051 4956 feature_gate.go:330] unrecognized feature gate: SignatureStores Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759060 4956 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759070 4956 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759083 4956 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759095 4956 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759107 4956 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759117 4956 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759126 4956 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759136 4956 feature_gate.go:330] unrecognized feature gate: OVNObservability Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759145 4956 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759154 4956 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759162 4956 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759169 4956 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759180 4956 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759190 4956 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759200 4956 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759209 4956 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759218 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759226 4956 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759234 4956 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759242 4956 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759250 4956 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759262 4956 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759272 4956 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759281 4956 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759292 4956 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759303 4956 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759314 4956 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759325 4956 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759340 4956 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759351 4956 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759361 4956 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759370 4956 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.759380 4956 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.759398 4956 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.759833 4956 server.go:940] "Client rotation is on, will bootstrap in background" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.768200 4956 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.768572 4956 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.771101 4956 server.go:997] "Starting client certificate rotation" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.771158 4956 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.772247 4956 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2025-12-30 00:53:23.552043844 +0000 UTC Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.772464 4956 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 800h0m48.779583831s for next certificate rotation Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.797383 4956 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.800947 4956 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.824089 4956 log.go:25] "Validated CRI v1 runtime API" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.855094 4956 log.go:25] "Validated CRI v1 image API" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.857460 4956 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.865427 4956 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-11-26-16-49-05-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.865472 4956 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.889931 4956 manager.go:217] Machine: {Timestamp:2025-11-26 16:52:34.886085805 +0000 UTC m=+0.582046427 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2800000 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:4a4a5395-957e-4512-a554-c3c322e283ff BootID:fcb74053-066d-43b8-98e1-af830e827a8c Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:a7:f6:30 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:a7:f6:30 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:cf:c9:ff Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:90:4c:67 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:8d:eb:7c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:12:b1:65 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:66:f9:1f:fe:09:e1 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:a6:8b:9e:70:ee:be Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.890393 4956 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.890808 4956 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.891608 4956 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.892090 4956 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.892150 4956 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.892494 4956 topology_manager.go:138] "Creating topology manager with none policy" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.892514 4956 container_manager_linux.go:303] "Creating device plugin manager" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.893052 4956 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.893102 4956 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.893425 4956 state_mem.go:36] "Initialized new in-memory state store" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.893568 4956 server.go:1245] "Using root directory" path="/var/lib/kubelet" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.897953 4956 kubelet.go:418] "Attempting to sync node with API server" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.897992 4956 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.898033 4956 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.898064 4956 kubelet.go:324] "Adding apiserver pod source" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.898085 4956 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.903945 4956 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.905244 4956 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.910007 4956 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.910545 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.912063 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.912554 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.912270 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.913893 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914582 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914622 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914639 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914672 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914691 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914709 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914736 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914760 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914778 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914809 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.914824 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.916033 4956 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.917465 4956 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.917678 4956 server.go:1280] "Started kubelet" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.919025 4956 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.919196 4956 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Nov 26 16:52:34 crc systemd[1]: Started Kubernetes Kubelet. Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.920471 4956 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.921764 4956 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.921823 4956 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.921977 4956 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 10:26:34.248722449 +0000 UTC Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.922396 4956 volume_manager.go:287] "The desired_state_of_world populator starts" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.922421 4956 volume_manager.go:289] "Starting Kubelet Volume Manager" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.922646 4956 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.922632 4956 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.923400 4956 server.go:460] "Adding debug handlers to kubelet server" Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.923539 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.923633 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.924953 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932669 4956 factory.go:153] Registering CRI-O factory Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932708 4956 factory.go:221] Registration of the crio container factory successfully Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932800 4956 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932817 4956 factory.go:55] Registering systemd factory Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932829 4956 factory.go:221] Registration of the systemd container factory successfully Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932862 4956 factory.go:103] Registering Raw factory Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.932911 4956 manager.go:1196] Started watching for new ooms in manager Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933342 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933399 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933417 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933436 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933454 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933473 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933489 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933506 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933526 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933544 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933561 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933580 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933597 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933618 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933635 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933651 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933670 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933737 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933795 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933815 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933834 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933855 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933891 4956 manager.go:319] Starting recovery of all containers Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933904 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.933998 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934020 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934038 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934059 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934080 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934099 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934140 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934158 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934175 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934193 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934260 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934289 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934304 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934363 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934379 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934394 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934411 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934431 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934448 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934463 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934562 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934585 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934602 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934620 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934646 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934665 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934679 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934738 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934753 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934774 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934794 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934812 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934828 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934915 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934935 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934950 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.934993 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935012 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935028 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935088 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935111 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935128 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935142 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935159 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935176 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935192 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935208 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935270 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935288 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935306 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935321 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935335 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935351 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935409 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935430 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935445 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935464 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935482 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935502 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.935574 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.939097 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.939316 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.939339 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.939367 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.939417 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.939434 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946101 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946206 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946236 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946262 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946281 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946373 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946398 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946415 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946481 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946507 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946529 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946548 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946561 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946578 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.946593 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947065 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947098 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947116 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947136 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947246 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947411 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.946238 4956 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.187b9caf1a266728 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 16:52:34.9176154 +0000 UTC m=+0.613575982,LastTimestamp:2025-11-26 16:52:34.9176154 +0000 UTC m=+0.613575982,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947492 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947516 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947582 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947659 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947685 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947710 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.947956 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948096 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948193 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948225 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948344 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948384 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948402 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948427 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.948902 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949077 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949200 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949229 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949447 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949539 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949573 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949602 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949658 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949689 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949706 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.949722 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950112 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950148 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950173 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950186 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950199 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950222 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950297 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950374 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950449 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950464 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950482 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950587 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950608 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950629 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950650 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950665 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950781 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950926 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.950995 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951016 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951307 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951379 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951412 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951431 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951451 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951474 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951491 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951512 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951527 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951544 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951564 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.951583 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955785 4956 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955818 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955833 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955877 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955896 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955910 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955932 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.955990 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956010 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956023 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956036 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956052 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956064 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956078 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956091 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956105 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956120 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956132 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956149 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956160 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956173 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956192 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956205 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956220 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956230 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956241 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956254 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956264 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956277 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956288 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956300 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956313 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956349 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956360 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956375 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956385 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956397 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956408 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956417 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956442 4956 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956451 4956 reconstruct.go:97] "Volume reconstruction finished" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.956458 4956 reconciler.go:26] "Reconciler: start to sync state" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.973988 4956 manager.go:324] Recovery completed Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.990247 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.990375 4956 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.993015 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.993101 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.993136 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.993853 4956 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.993963 4956 status_manager.go:217] "Starting to sync pod status with apiserver" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.994010 4956 kubelet.go:2335] "Starting kubelet main sync loop" Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.994091 4956 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Nov 26 16:52:34 crc kubenswrapper[4956]: W1126 16:52:34.994580 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:34 crc kubenswrapper[4956]: E1126 16:52:34.994630 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.994670 4956 cpu_manager.go:225] "Starting CPU manager" policy="none" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.994729 4956 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Nov 26 16:52:34 crc kubenswrapper[4956]: I1126 16:52:34.994755 4956 state_mem.go:36] "Initialized new in-memory state store" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.018355 4956 policy_none.go:49] "None policy: Start" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.019227 4956 memory_manager.go:170] "Starting memorymanager" policy="None" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.019263 4956 state_mem.go:35] "Initializing new in-memory state store" Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.023546 4956 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.092494 4956 manager.go:334] "Starting Device Plugin manager" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.092581 4956 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.092603 4956 server.go:79] "Starting device plugin registration server" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.093303 4956 eviction_manager.go:189] "Eviction manager: starting control loop" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.093332 4956 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.093556 4956 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.093901 4956 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.093924 4956 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.094201 4956 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.094405 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.097127 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.097194 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.097222 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.097567 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.098013 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.098189 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.099421 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.099504 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.099540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.099760 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100019 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100083 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100351 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100490 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100840 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100955 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.100998 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.101018 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.101228 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.101740 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.101810 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103069 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103095 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103106 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103231 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103260 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103278 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.103506 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.104275 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.104362 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.105582 4956 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.105990 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.106014 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.106026 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.106242 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.106274 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.107077 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.107116 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.107134 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.107424 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.107442 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.107453 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.108452 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.108474 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.108485 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.125843 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.158555 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.158627 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.158675 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.158717 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.158820 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.158941 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159002 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159049 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159093 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159136 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159181 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159250 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159291 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159330 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.159358 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.193943 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.195594 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.195675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.195693 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.195784 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.196638 4956 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260301 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260421 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260489 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260526 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260563 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260606 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260632 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260652 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260729 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260736 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260822 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260787 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260833 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260908 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260965 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261043 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260958 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.260910 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261227 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261352 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261469 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261514 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261563 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261639 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261703 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261611 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261745 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261821 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261931 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.261943 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.397660 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.399745 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.399817 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.399831 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.399916 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.400621 4956 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.437431 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.447656 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.466148 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.494220 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: W1126 16:52:35.497802 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-1473bde64fea3f0c296a19a4b5e0ccc7aa4b8359cf44eafbed1c9e3fe1c5439b WatchSource:0}: Error finding container 1473bde64fea3f0c296a19a4b5e0ccc7aa4b8359cf44eafbed1c9e3fe1c5439b: Status 404 returned error can't find the container with id 1473bde64fea3f0c296a19a4b5e0ccc7aa4b8359cf44eafbed1c9e3fe1c5439b Nov 26 16:52:35 crc kubenswrapper[4956]: W1126 16:52:35.503981 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-7de382a60fbaef3df4c0f417ac3773c59c754d5e3242ae522d14202b199368b1 WatchSource:0}: Error finding container 7de382a60fbaef3df4c0f417ac3773c59c754d5e3242ae522d14202b199368b1: Status 404 returned error can't find the container with id 7de382a60fbaef3df4c0f417ac3773c59c754d5e3242ae522d14202b199368b1 Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.504598 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Nov 26 16:52:35 crc kubenswrapper[4956]: W1126 16:52:35.505749 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-33ec90d162456f1e1137cece368812c72e3823e42556294d8592000499e78819 WatchSource:0}: Error finding container 33ec90d162456f1e1137cece368812c72e3823e42556294d8592000499e78819: Status 404 returned error can't find the container with id 33ec90d162456f1e1137cece368812c72e3823e42556294d8592000499e78819 Nov 26 16:52:35 crc kubenswrapper[4956]: W1126 16:52:35.520304 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-0c79345693b2dc2f2aae8a4f078fdc4231f6578409bf8a46656a4fa4d2ae5e5e WatchSource:0}: Error finding container 0c79345693b2dc2f2aae8a4f078fdc4231f6578409bf8a46656a4fa4d2ae5e5e: Status 404 returned error can't find the container with id 0c79345693b2dc2f2aae8a4f078fdc4231f6578409bf8a46656a4fa4d2ae5e5e Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.527481 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Nov 26 16:52:35 crc kubenswrapper[4956]: W1126 16:52:35.728553 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.729238 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:35 crc kubenswrapper[4956]: W1126 16:52:35.779586 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.779710 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.801280 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.803185 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.803270 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.803290 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.803342 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:52:35 crc kubenswrapper[4956]: E1126 16:52:35.804238 4956 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.918764 4956 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:35 crc kubenswrapper[4956]: I1126 16:52:35.922815 4956 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-11-18 13:36:27.734084706 +0000 UTC Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.000881 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"33ec90d162456f1e1137cece368812c72e3823e42556294d8592000499e78819"} Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.002614 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"7de382a60fbaef3df4c0f417ac3773c59c754d5e3242ae522d14202b199368b1"} Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.004351 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"1473bde64fea3f0c296a19a4b5e0ccc7aa4b8359cf44eafbed1c9e3fe1c5439b"} Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.005787 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"0c79345693b2dc2f2aae8a4f078fdc4231f6578409bf8a46656a4fa4d2ae5e5e"} Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.007170 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f5e982e55f826bdf3ea6c621185255b6a6ac532b0a18e0d884da449df3421c8e"} Nov 26 16:52:36 crc kubenswrapper[4956]: W1126 16:52:36.040182 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:36 crc kubenswrapper[4956]: E1126 16:52:36.040263 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:36 crc kubenswrapper[4956]: E1126 16:52:36.328842 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Nov 26 16:52:36 crc kubenswrapper[4956]: W1126 16:52:36.481265 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:36 crc kubenswrapper[4956]: E1126 16:52:36.481385 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.605230 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.608850 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.608937 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.608958 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.609003 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:52:36 crc kubenswrapper[4956]: E1126 16:52:36.609795 4956 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.918977 4956 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.923046 4956 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-07 00:43:53.277507928 +0000 UTC Nov 26 16:52:36 crc kubenswrapper[4956]: I1126 16:52:36.923183 4956 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 247h51m16.354332931s for next certificate rotation Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.013299 4956 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5" exitCode=0 Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.013384 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.013494 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.015379 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.015421 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.015433 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.017988 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.018021 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.018034 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.020718 4956 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209" exitCode=0 Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.020773 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.020898 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.021943 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.021979 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.021990 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.023587 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18" exitCode=0 Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.023645 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.023723 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.024984 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.025047 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.025062 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.026204 4956 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="ec32df034cd0721ea842341e57af6a4cef5749f5806a5fc205657705a1d53bfe" exitCode=0 Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.026238 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"ec32df034cd0721ea842341e57af6a4cef5749f5806a5fc205657705a1d53bfe"} Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.026328 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.027567 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.027624 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.027625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.028143 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.028721 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.028760 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.028773 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:37 crc kubenswrapper[4956]: I1126 16:52:37.919503 4956 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:37 crc kubenswrapper[4956]: E1126 16:52:37.930304 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.032539 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.032614 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.034277 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.034334 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.034361 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.037356 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.037542 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.039269 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.039315 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.039331 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.043097 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.043144 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.043161 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.043164 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.044418 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.044476 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.044491 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.048684 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.048743 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.048759 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.051247 4956 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="150a5dc03e41dbff21404b2faa19c1a1d9d056f4c0a494f62a0ecbe62d9e4b6b" exitCode=0 Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.051293 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"150a5dc03e41dbff21404b2faa19c1a1d9d056f4c0a494f62a0ecbe62d9e4b6b"} Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.051349 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.052170 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.052211 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.052222 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.210228 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.212565 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.212638 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.212653 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:38 crc kubenswrapper[4956]: I1126 16:52:38.212702 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:52:38 crc kubenswrapper[4956]: E1126 16:52:38.213375 4956 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.9:6443: connect: connection refused" node="crc" Nov 26 16:52:38 crc kubenswrapper[4956]: W1126 16:52:38.300688 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:38 crc kubenswrapper[4956]: E1126 16:52:38.300790 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:38 crc kubenswrapper[4956]: W1126 16:52:38.435999 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.9:6443: connect: connection refused Nov 26 16:52:38 crc kubenswrapper[4956]: E1126 16:52:38.436120 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.9:6443: connect: connection refused" logger="UnhandledError" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.058134 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227"} Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.058213 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6"} Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.058253 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.059503 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.059542 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.059563 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062243 4956 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="e2488af2383807750fb8b7f46a6e7c41ab56bc6ac71e5a596269535dfaee3db4" exitCode=0 Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062328 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062361 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062404 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062533 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062675 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.062737 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"e2488af2383807750fb8b7f46a6e7c41ab56bc6ac71e5a596269535dfaee3db4"} Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.063564 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.063589 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.063600 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064288 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064311 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064484 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064569 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064489 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.064621 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:39 crc kubenswrapper[4956]: I1126 16:52:39.854795 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.051344 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.071915 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"bfc47791126fac6852148d47372d4a71b81200777d0f31e2deb4ca074a99d37f"} Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.071995 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.071989 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"102ab92894b9404dacf77a7c54cf8053dba3370cafe1fd26634654db48779ed1"} Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.072102 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.072116 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"2f66ea5699c085042805e6c0645c749139ff6ee4c76aea681a4e15c5eecb2ede"} Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.073501 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.073551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.073570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.073650 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.073664 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.073672 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.088339 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.088526 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.089588 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.089660 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.089681 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.682058 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:40 crc kubenswrapper[4956]: I1126 16:52:40.810374 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.083376 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"0a4268dabea01571e3392efe8be957e50d25344115a017f9a62d21dc6ea516a9"} Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.083454 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b736669e536686fe85175ce17e588699dd801f18b0ad1019b2877766a1fe8f49"} Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.083497 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.083764 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.085087 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.085128 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.085140 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.085428 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.085490 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.085533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.413740 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.415273 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.415340 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.415359 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.415404 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.611023 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.611280 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.612901 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.612945 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:41 crc kubenswrapper[4956]: I1126 16:52:41.612954 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.086732 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.086761 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.088630 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.088712 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.088749 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.088981 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.089063 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.089087 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:42 crc kubenswrapper[4956]: I1126 16:52:42.523085 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.089317 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.091058 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.091129 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.091153 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.334518 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.334773 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.336416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.336472 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.336491 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:43 crc kubenswrapper[4956]: I1126 16:52:43.806026 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Nov 26 16:52:44 crc kubenswrapper[4956]: I1126 16:52:44.093955 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:44 crc kubenswrapper[4956]: I1126 16:52:44.095730 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:44 crc kubenswrapper[4956]: I1126 16:52:44.095777 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:44 crc kubenswrapper[4956]: I1126 16:52:44.095796 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:45 crc kubenswrapper[4956]: E1126 16:52:45.105906 4956 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.335528 4956 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.335650 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.698672 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.699049 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.701833 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.701926 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.701948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:46 crc kubenswrapper[4956]: I1126 16:52:46.706257 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:47 crc kubenswrapper[4956]: I1126 16:52:47.104368 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:47 crc kubenswrapper[4956]: I1126 16:52:47.105949 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:47 crc kubenswrapper[4956]: I1126 16:52:47.106003 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:47 crc kubenswrapper[4956]: I1126 16:52:47.106017 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:47 crc kubenswrapper[4956]: I1126 16:52:47.111334 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:48 crc kubenswrapper[4956]: I1126 16:52:48.107245 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:48 crc kubenswrapper[4956]: I1126 16:52:48.108738 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:48 crc kubenswrapper[4956]: I1126 16:52:48.108898 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:48 crc kubenswrapper[4956]: I1126 16:52:48.108928 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:48 crc kubenswrapper[4956]: I1126 16:52:48.919934 4956 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Nov 26 16:52:49 crc kubenswrapper[4956]: W1126 16:52:49.049389 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 26 16:52:49 crc kubenswrapper[4956]: I1126 16:52:49.049601 4956 trace.go:236] Trace[394887487]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 16:52:39.046) (total time: 10003ms): Nov 26 16:52:49 crc kubenswrapper[4956]: Trace[394887487]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10002ms (16:52:49.049) Nov 26 16:52:49 crc kubenswrapper[4956]: Trace[394887487]: [10.003060541s] [10.003060541s] END Nov 26 16:52:49 crc kubenswrapper[4956]: E1126 16:52:49.049645 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 26 16:52:49 crc kubenswrapper[4956]: W1126 16:52:49.498100 4956 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout Nov 26 16:52:49 crc kubenswrapper[4956]: I1126 16:52:49.498259 4956 trace.go:236] Trace[396552747]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 16:52:39.496) (total time: 10001ms): Nov 26 16:52:49 crc kubenswrapper[4956]: Trace[396552747]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (16:52:49.498) Nov 26 16:52:49 crc kubenswrapper[4956]: Trace[396552747]: [10.001788066s] [10.001788066s] END Nov 26 16:52:49 crc kubenswrapper[4956]: E1126 16:52:49.498303 4956 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Nov 26 16:52:49 crc kubenswrapper[4956]: I1126 16:52:49.763361 4956 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Nov 26 16:52:49 crc kubenswrapper[4956]: I1126 16:52:49.763451 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 26 16:52:49 crc kubenswrapper[4956]: I1126 16:52:49.772039 4956 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\": RBAC: [clusterrole.rbac.authorization.k8s.io \"system:openshift:public-info-viewer\" not found, clusterrole.rbac.authorization.k8s.io \"system:public-info-viewer\" not found]","reason":"Forbidden","details":{},"code":403} Nov 26 16:52:49 crc kubenswrapper[4956]: I1126 16:52:49.772108 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Nov 26 16:52:50 crc kubenswrapper[4956]: I1126 16:52:50.823658 4956 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]log ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]etcd ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-apiserver-admission-initializer ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/openshift.io-api-request-count-filter ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/openshift.io-startkubeinformers ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/openshift.io-openshift-apiserver-reachable ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/openshift.io-oauth-apiserver-reachable ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/generic-apiserver-start-informers ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/priority-and-fairness-config-consumer ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/priority-and-fairness-filter ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/storage-object-count-tracker-hook ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-apiextensions-informers ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-apiextensions-controllers ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/crd-informer-synced ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-system-namespaces-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-cluster-authentication-info-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-kube-apiserver-identity-lease-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-kube-apiserver-identity-lease-garbage-collector ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-legacy-token-tracking-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-service-ip-repair-controllers ok Nov 26 16:52:50 crc kubenswrapper[4956]: [-]poststarthook/rbac/bootstrap-roles failed: reason withheld Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/scheduling/bootstrap-system-priority-classes ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/priority-and-fairness-config-producer ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/bootstrap-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/aggregator-reload-proxy-client-cert ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/start-kube-aggregator-informers ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-status-local-available-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-status-remote-available-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-registration-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-wait-for-first-sync ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-discovery-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/kube-apiserver-autoregistration ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]autoregister-completion ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-openapi-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: [+]poststarthook/apiservice-openapiv3-controller ok Nov 26 16:52:50 crc kubenswrapper[4956]: livez check failed Nov 26 16:52:50 crc kubenswrapper[4956]: I1126 16:52:50.823727 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:52:53 crc kubenswrapper[4956]: I1126 16:52:53.839508 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Nov 26 16:52:53 crc kubenswrapper[4956]: I1126 16:52:53.839747 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:53 crc kubenswrapper[4956]: I1126 16:52:53.841114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:53 crc kubenswrapper[4956]: I1126 16:52:53.841151 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:53 crc kubenswrapper[4956]: I1126 16:52:53.841159 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:53 crc kubenswrapper[4956]: I1126 16:52:53.854560 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.009855 4956 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.124429 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.125854 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.125949 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.125971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.715259 4956 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.759064 4956 trace.go:236] Trace[1508396072]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 16:52:41.732) (total time: 13026ms): Nov 26 16:52:54 crc kubenswrapper[4956]: Trace[1508396072]: ---"Objects listed" error: 13025ms (16:52:54.758) Nov 26 16:52:54 crc kubenswrapper[4956]: Trace[1508396072]: [13.026142512s] [13.026142512s] END Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.759164 4956 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.759466 4956 trace.go:236] Trace[829306756]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (26-Nov-2025 16:52:43.285) (total time: 11473ms): Nov 26 16:52:54 crc kubenswrapper[4956]: Trace[829306756]: ---"Objects listed" error: 11473ms (16:52:54.759) Nov 26 16:52:54 crc kubenswrapper[4956]: Trace[829306756]: [11.473746375s] [11.473746375s] END Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.759503 4956 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 26 16:52:54 crc kubenswrapper[4956]: E1126 16:52:54.760341 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.761132 4956 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Nov 26 16:52:54 crc kubenswrapper[4956]: E1126 16:52:54.764618 4956 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.814438 4956 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56404->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.814536 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56404->192.168.126.11:17697: read: connection reset by peer" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.814439 4956 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56408->192.168.126.11:17697: read: connection reset by peer" start-of-body= Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.814738 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:56408->192.168.126.11:17697: read: connection reset by peer" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.909922 4956 apiserver.go:52] "Watching apiserver" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.914446 4956 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.914821 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.915332 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.915464 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.915567 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.915573 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:54 crc kubenswrapper[4956]: E1126 16:52:54.915686 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.915792 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:54 crc kubenswrapper[4956]: E1126 16:52:54.915926 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.916184 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:54 crc kubenswrapper[4956]: E1126 16:52:54.916272 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.920768 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.920918 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921090 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921139 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921227 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921303 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921446 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921448 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.921596 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.922552 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.923525 4956 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.927515 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.944987 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.960665 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.962859 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.962941 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.962986 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963030 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963068 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963114 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963148 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963207 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963283 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963321 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963357 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963395 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963426 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963458 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963492 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963523 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963558 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963660 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963685 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963700 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.963766 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964143 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964187 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964220 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964255 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964295 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964218 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964430 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964456 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964484 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964493 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964627 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.964690 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.965072 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.965495 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.965614 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.965651 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.965655 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966366 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966406 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966432 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966477 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966668 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966947 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967029 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967051 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967175 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.966854 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967427 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967696 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967755 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967772 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967809 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967844 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967900 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.967943 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.968006 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.968438 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.968520 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.968556 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.968822 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.968943 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969095 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969134 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969164 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969180 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969338 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969387 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969428 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969465 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969608 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969670 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969709 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969746 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969780 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969837 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969918 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969952 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969987 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970023 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970090 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970266 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970365 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970399 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970438 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970477 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970511 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970550 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970589 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970634 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970667 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970710 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.969618 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970142 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970075 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970386 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970495 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.971348 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.971569 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.971604 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.971809 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.971972 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972047 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.970753 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972408 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972444 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972480 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972537 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972596 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 16:52:54 crc kubenswrapper[4956]: E1126 16:52:54.972641 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:52:55.472608086 +0000 UTC m=+21.168568668 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972685 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972732 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972766 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972801 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972838 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972938 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972955 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.972979 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973018 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973056 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973091 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973133 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973170 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973203 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973224 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973240 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973281 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973320 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973352 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973386 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973424 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973442 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973460 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973462 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973506 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973539 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973765 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974102 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.973767 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974207 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974242 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974273 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974302 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974331 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974357 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974383 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974409 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974436 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974463 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974487 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974513 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974537 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974561 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974586 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974610 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974633 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974657 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974679 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974701 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974724 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974746 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974774 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974797 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974799 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974818 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974843 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974888 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974916 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974940 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974965 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.974985 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.975007 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.976695 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.975043 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.976816 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.976880 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.977281 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978165 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978160 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978219 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978499 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978794 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978802 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.978884 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.979114 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.979592 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.979647 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.979826 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.979856 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.980121 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.982268 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.982410 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.983219 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984115 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984265 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984577 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984663 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984693 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984710 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984806 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.985196 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.985291 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.984915 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.985610 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.985763 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.985795 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.985980 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986030 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986024 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986231 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986360 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986076 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986421 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.986600 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987244 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987373 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987409 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987385 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987538 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987560 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987887 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.987853 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.988006 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.988594 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.989004 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.989451 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.989513 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.989688 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.990130 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.990220 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.990724 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.991127 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.991253 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.991515 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.992473 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.993082 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994696 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994741 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994768 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994790 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994813 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994835 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994904 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994938 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994961 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.994981 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995005 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995027 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995049 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995074 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995098 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995185 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995237 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995260 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995310 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995333 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995361 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995409 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995431 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995477 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995520 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995570 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995592 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995642 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995669 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995692 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995738 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995761 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995808 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Nov 26 16:52:54 crc kubenswrapper[4956]: I1126 16:52:54.995830 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.995853 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001098 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001150 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001189 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001223 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001257 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001296 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001348 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001382 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001419 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001505 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001602 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001650 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001687 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001736 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001783 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001817 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001852 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.996238 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.997438 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.997520 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.997792 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.998206 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.999142 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.999160 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.999527 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:54.999749 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.000133 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.000756 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.000984 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001033 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.002555 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.003065 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.003138 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.003327 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.003557 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.004330 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.005324 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.005780 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.005938 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.006931 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.007170 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.007643 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.007228 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.007559 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.001917 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.007941 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.008044 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.008058 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.008216 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.008300 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.009325 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.009614 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010005 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010087 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010388 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010424 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010458 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010490 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010521 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010559 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010593 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010625 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010655 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010689 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010725 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010759 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010793 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010822 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010846 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010909 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010959 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.010995 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011028 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011063 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011096 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011131 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011166 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011402 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011447 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011484 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011529 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011565 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011603 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011648 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011683 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011718 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011752 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011799 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011833 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011908 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.011952 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012249 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012276 4956 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012296 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012315 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012335 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012354 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012372 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012392 4956 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012410 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012429 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012448 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012502 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012522 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012540 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012559 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012577 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012595 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012612 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012639 4956 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012657 4956 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012674 4956 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012695 4956 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012713 4956 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012728 4956 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012745 4956 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012763 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012780 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012858 4956 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012946 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012964 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012982 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.012999 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013016 4956 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013032 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013049 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013064 4956 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013260 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013277 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013294 4956 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013310 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013330 4956 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013348 4956 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013365 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013381 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013329 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013397 4956 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013622 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013718 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013721 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013800 4956 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013839 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013937 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.013998 4956 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014029 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014059 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014094 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014127 4956 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014154 4956 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014181 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014207 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014234 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014261 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014287 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014314 4956 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014342 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014766 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014943 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014963 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.014991 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.016139 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017457 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.015752 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017568 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017624 4956 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017735 4956 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.016693 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.018715 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017061 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.016165 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.016238 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.016584 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017065 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.017378 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019105 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019337 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019403 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019423 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019674 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019733 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.019766 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.020301 4956 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.020858 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.020004 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.024846 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:55.524816028 +0000 UTC m=+21.220776610 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.020476 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.020604 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.022164 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.025495 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:55.525475556 +0000 UTC m=+21.221436138 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.022931 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.024594 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.025820 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.025853 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.027999 4956 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028024 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028043 4956 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028090 4956 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028109 4956 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028127 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028147 4956 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028168 4956 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028192 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028212 4956 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028230 4956 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028248 4956 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028266 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028284 4956 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028302 4956 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028319 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028337 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028355 4956 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028375 4956 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028426 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028474 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028493 4956 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028511 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028527 4956 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028542 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028557 4956 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028574 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028589 4956 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028605 4956 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028621 4956 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.028637 4956 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.031998 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032040 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032084 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032098 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032112 4956 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032127 4956 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032175 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032188 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032200 4956 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032215 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032247 4956 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032263 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032275 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032288 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032301 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032332 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032345 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032359 4956 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032372 4956 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032405 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032418 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032430 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032460 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032510 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032527 4956 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032543 4956 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032587 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032605 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032623 4956 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032639 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032678 4956 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032698 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032712 4956 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032755 4956 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032771 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032786 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032802 4956 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032837 4956 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032900 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032920 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032937 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032952 4956 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.032996 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.033014 4956 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.033066 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.033474 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.036489 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.036682 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.038191 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.039014 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.039158 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.040148 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.039497 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.039506 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.040125 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.041654 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.042263 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.042643 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.042830 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.043548 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.043592 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.043605 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.043692 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:55.543671332 +0000 UTC m=+21.239631884 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.044317 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.045435 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.045797 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.045511 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.046164 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.048212 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.048238 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.048252 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.048311 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:55.54828681 +0000 UTC m=+21.244247352 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.048533 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.048640 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.055095 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.056730 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.057355 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.059113 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.059371 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.063304 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.063598 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.064089 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.066199 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.066286 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.067023 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.067339 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.068729 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.072919 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.074745 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.075146 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.074897 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.077497 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.078689 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.079895 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.084082 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.085040 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.085894 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.087746 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.087891 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.087846 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.088650 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.091102 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.094917 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.096812 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.096888 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.096913 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.098479 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.101626 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.103042 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.104254 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.105853 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.107232 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.108153 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.109501 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.114329 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.114837 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.115834 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.116336 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.117450 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.117592 4956 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.117709 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.119598 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.120203 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.121062 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.122641 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.123437 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.124713 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.125340 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.127125 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.127108 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.127807 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.129149 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.129883 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.130140 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.131121 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.132116 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.133443 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.134096 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6" exitCode=255 Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.134634 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.135916 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136004 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136123 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136141 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136148 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136330 4956 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136380 4956 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136418 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136433 4956 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136444 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136454 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136464 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136475 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136485 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136495 4956 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136504 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136774 4956 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136791 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136820 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136852 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136887 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136902 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136917 4956 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136934 4956 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136947 4956 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136961 4956 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.136974 4956 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137767 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137784 4956 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137799 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137848 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137892 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137907 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137920 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137933 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137946 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137960 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137974 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137986 4956 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.137999 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138011 4956 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138023 4956 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138035 4956 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138047 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138059 4956 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138071 4956 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138083 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138094 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138106 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138117 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.138443 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.139179 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.139565 4956 kubelet.go:1929] "Failed creating a mirror pod for" err="pods \"kube-controller-manager-crc\" already exists" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.140578 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.140806 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.141536 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.142693 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.144648 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6"} Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.144748 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.150849 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.155280 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.155888 4956 scope.go:117] "RemoveContainer" containerID="a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.161620 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.171074 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.181019 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.191308 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.200033 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.209844 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.219490 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.228354 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.238324 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.238799 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.250383 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Nov 26 16:52:55 crc kubenswrapper[4956]: W1126 16:52:55.250681 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-ad0cb5f9ae7bd1e03cfda5e61ec38e367d63b68f77e7b0ca2803858a224e0226 WatchSource:0}: Error finding container ad0cb5f9ae7bd1e03cfda5e61ec38e367d63b68f77e7b0ca2803858a224e0226: Status 404 returned error can't find the container with id ad0cb5f9ae7bd1e03cfda5e61ec38e367d63b68f77e7b0ca2803858a224e0226 Nov 26 16:52:55 crc kubenswrapper[4956]: W1126 16:52:55.261627 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-3e2e90f0b5687308e952729e98639a533dfd9b343dd9805f1740c17f4642320a WatchSource:0}: Error finding container 3e2e90f0b5687308e952729e98639a533dfd9b343dd9805f1740c17f4642320a: Status 404 returned error can't find the container with id 3e2e90f0b5687308e952729e98639a533dfd9b343dd9805f1740c17f4642320a Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.265050 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.551567 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.551680 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.551718 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.551742 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.551766 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.551839 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:52:56.551812416 +0000 UTC m=+22.247772968 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.551927 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.551989 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:56.5519798 +0000 UTC m=+22.247940352 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.551988 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.551991 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552032 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:56.552022781 +0000 UTC m=+22.247983333 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552038 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552061 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552137 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:56.552110534 +0000 UTC m=+22.248071126 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552158 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552171 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552181 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: E1126 16:52:55.552209 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:56.552200086 +0000 UTC m=+22.248160638 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.815506 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.833943 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.852992 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.863970 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.872413 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.881024 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.891456 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.902062 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:55 crc kubenswrapper[4956]: I1126 16:52:55.911733 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.141136 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.141603 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"ad0cb5f9ae7bd1e03cfda5e61ec38e367d63b68f77e7b0ca2803858a224e0226"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.151983 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.154347 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.154843 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.156005 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"65cc40fbccc57e50dac00d9c08c39dcd9880e5d4fa5d741a78fe3bc7f1f3cc5b"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.158428 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.158459 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.158469 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"3e2e90f0b5687308e952729e98639a533dfd9b343dd9805f1740c17f4642320a"} Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.164731 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.171678 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.184184 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.203169 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.215072 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.225651 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.252885 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.281415 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.300421 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.318125 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.338744 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.352151 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.366049 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.381124 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.395143 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.414328 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.433249 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:56Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.562205 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.562265 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.562299 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.562320 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.562343 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562475 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562529 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562546 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562594 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562653 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:52:58.56261661 +0000 UTC m=+24.258577162 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562475 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562607 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562723 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562740 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562753 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:58.562712772 +0000 UTC m=+24.258673364 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562786 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:58.562768294 +0000 UTC m=+24.258728846 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562807 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:58.562800515 +0000 UTC m=+24.258761067 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.562821 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:52:58.562814805 +0000 UTC m=+24.258775347 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.994627 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.994628 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.994754 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.994811 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:52:56 crc kubenswrapper[4956]: I1126 16:52:56.994645 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:56 crc kubenswrapper[4956]: E1126 16:52:56.994907 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.579491 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.579656 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.579754 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:53:02.579706541 +0000 UTC m=+28.275667103 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.579785 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.579834 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.579887 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:02.579838744 +0000 UTC m=+28.275799366 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.579925 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.579967 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.579999 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580017 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580032 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580077 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:02.580064401 +0000 UTC m=+28.276024963 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580084 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580106 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580122 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580168 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:02.580152473 +0000 UTC m=+28.276113085 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580216 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.580259 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:02.580242746 +0000 UTC m=+28.276203398 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.994485 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.994925 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.994767 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.995680 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:52:58 crc kubenswrapper[4956]: I1126 16:52:58.996158 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:52:58 crc kubenswrapper[4956]: E1126 16:52:58.996416 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.168428 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96"} Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.192643 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.213343 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.233966 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.253412 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.271319 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.288948 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.305215 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:52:59 crc kubenswrapper[4956]: I1126 16:52:59.339207 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:52:59Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:00 crc kubenswrapper[4956]: I1126 16:53:00.995117 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:00 crc kubenswrapper[4956]: I1126 16:53:00.995145 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:00 crc kubenswrapper[4956]: I1126 16:53:00.995186 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:00 crc kubenswrapper[4956]: E1126 16:53:00.995256 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:00 crc kubenswrapper[4956]: E1126 16:53:00.995401 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:00 crc kubenswrapper[4956]: E1126 16:53:00.995519 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.165545 4956 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.167458 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.167492 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.167503 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.167562 4956 kubelet_node_status.go:76] "Attempting to register node" node="crc" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.174198 4956 kubelet_node_status.go:115] "Node was previously registered" node="crc" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.174531 4956 kubelet_node_status.go:79] "Successfully registered node" node="crc" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.175749 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.175787 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.175796 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.175810 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.175820 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.190962 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.195812 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.195880 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.195897 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.195913 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.195923 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.206274 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.209937 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.209989 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.210000 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.210017 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.210027 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.221458 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.224341 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.224378 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.224388 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.224402 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.224412 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.236940 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.241842 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.242021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.242273 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.242654 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.242757 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.245025 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-vhvtw"] Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.245636 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.252024 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 26 16:53:01 crc kubenswrapper[4956]: W1126 16:53:01.252221 4956 reflector.go:561] object-"openshift-dns"/"node-resolver-dockercfg-kz9s7": failed to list *v1.Secret: secrets "node-resolver-dockercfg-kz9s7" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-dns": no relationship found between node 'crc' and this object Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.252276 4956 reflector.go:158] "Unhandled Error" err="object-\"openshift-dns\"/\"node-resolver-dockercfg-kz9s7\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"node-resolver-dockercfg-kz9s7\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-dns\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.252292 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-9rvrb"] Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.252695 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.253136 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.254823 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.257539 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.257796 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.258193 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.262661 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: E1126 16:53:01.262778 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.265495 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.265610 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.265690 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.265757 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.265826 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.272088 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.273581 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.284200 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.298315 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.306800 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-proxy-tls\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.306845 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8dtp\" (UniqueName: \"kubernetes.io/projected/adc2efe4-3ec6-4e5f-8dd8-76954150acb1-kube-api-access-s8dtp\") pod \"node-resolver-vhvtw\" (UID: \"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\") " pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.306880 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-rootfs\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.306900 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jrmtl\" (UniqueName: \"kubernetes.io/projected/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-kube-api-access-jrmtl\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.307069 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/adc2efe4-3ec6-4e5f-8dd8-76954150acb1-hosts-file\") pod \"node-resolver-vhvtw\" (UID: \"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\") " pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.307121 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-mcd-auth-proxy-config\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.314514 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.329750 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.340135 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.348528 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.360675 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.367570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.367614 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.367626 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.367642 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.367653 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.375398 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.390923 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.404414 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407618 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-proxy-tls\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407660 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8dtp\" (UniqueName: \"kubernetes.io/projected/adc2efe4-3ec6-4e5f-8dd8-76954150acb1-kube-api-access-s8dtp\") pod \"node-resolver-vhvtw\" (UID: \"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\") " pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407679 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-rootfs\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407697 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jrmtl\" (UniqueName: \"kubernetes.io/projected/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-kube-api-access-jrmtl\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407722 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/adc2efe4-3ec6-4e5f-8dd8-76954150acb1-hosts-file\") pod \"node-resolver-vhvtw\" (UID: \"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\") " pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407738 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-mcd-auth-proxy-config\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407925 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-rootfs\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.407990 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/adc2efe4-3ec6-4e5f-8dd8-76954150acb1-hosts-file\") pod \"node-resolver-vhvtw\" (UID: \"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\") " pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.408418 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-mcd-auth-proxy-config\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.419269 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-proxy-tls\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.424757 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.431925 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8dtp\" (UniqueName: \"kubernetes.io/projected/adc2efe4-3ec6-4e5f-8dd8-76954150acb1-kube-api-access-s8dtp\") pod \"node-resolver-vhvtw\" (UID: \"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\") " pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.433470 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jrmtl\" (UniqueName: \"kubernetes.io/projected/fadaf3cf-cfc5-4f60-bd4a-4eae814da018-kube-api-access-jrmtl\") pod \"machine-config-daemon-9rvrb\" (UID: \"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\") " pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.447508 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.462077 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.471358 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.471416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.471429 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.471448 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.471464 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.483446 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.500034 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.510120 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.526706 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.538788 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.574327 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.574652 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.574721 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.574810 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.574902 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.576501 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.631502 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-vz5zh"] Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.632128 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-m4dm4"] Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.632371 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.632660 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635030 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635059 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635110 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635535 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635582 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635728 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.635925 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.654500 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.674022 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.677989 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.678033 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.678046 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.678067 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.678083 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.705541 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710678 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-multus-certs\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710732 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710756 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-k8s-cni-cncf-io\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710782 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d94ea852-698d-41f1-93e4-df0322a23b47-cni-binary-copy\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710803 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/80ac0050-7f55-4845-ace7-a89f2442f884-multus-daemon-config\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710822 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-etc-kubernetes\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710842 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-cni-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710879 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-cnibin\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710902 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d94ea852-698d-41f1-93e4-df0322a23b47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710924 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxgn8\" (UniqueName: \"kubernetes.io/projected/d94ea852-698d-41f1-93e4-df0322a23b47-kube-api-access-bxgn8\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710958 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-os-release\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.710989 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-system-cni-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711015 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-socket-dir-parent\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711038 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-netns\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711059 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-hostroot\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711080 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-system-cni-dir\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711100 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-cnibin\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711122 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/80ac0050-7f55-4845-ace7-a89f2442f884-cni-binary-copy\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711143 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-cni-multus\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711173 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-os-release\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711194 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-cni-bin\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711213 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-85hhq\" (UniqueName: \"kubernetes.io/projected/80ac0050-7f55-4845-ace7-a89f2442f884-kube-api-access-85hhq\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711250 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-kubelet\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.711272 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-conf-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.728008 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.758963 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.772415 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.780681 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.780733 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.780750 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.780771 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.780785 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.788423 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.802271 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811827 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-multus-certs\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811890 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811911 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-k8s-cni-cncf-io\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811932 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d94ea852-698d-41f1-93e4-df0322a23b47-cni-binary-copy\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811950 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/80ac0050-7f55-4845-ace7-a89f2442f884-multus-daemon-config\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811972 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-etc-kubernetes\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.811991 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-cni-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812008 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-cnibin\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812025 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d94ea852-698d-41f1-93e4-df0322a23b47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812043 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxgn8\" (UniqueName: \"kubernetes.io/projected/d94ea852-698d-41f1-93e4-df0322a23b47-kube-api-access-bxgn8\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812070 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-os-release\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812097 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-system-cni-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812126 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-socket-dir-parent\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812152 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-netns\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812178 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-hostroot\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812203 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-system-cni-dir\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812225 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-cnibin\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812246 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/80ac0050-7f55-4845-ace7-a89f2442f884-cni-binary-copy\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812269 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-cni-multus\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812298 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-os-release\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812329 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-cni-bin\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812349 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-85hhq\" (UniqueName: \"kubernetes.io/projected/80ac0050-7f55-4845-ace7-a89f2442f884-kube-api-access-85hhq\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812373 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-kubelet\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812392 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-conf-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812465 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-conf-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812511 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-multus-certs\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812752 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-netns\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812836 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-hostroot\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812826 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-cnibin\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812905 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-os-release\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813003 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-cnibin\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813021 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-os-release\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813048 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-cni-bin\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812936 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-cni-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813030 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-run-k8s-cni-cncf-io\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812942 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-system-cni-dir\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813103 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-cni-multus\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.812948 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-etc-kubernetes\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813124 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-host-var-lib-kubelet\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813160 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-system-cni-dir\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813203 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/80ac0050-7f55-4845-ace7-a89f2442f884-multus-socket-dir-parent\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.813316 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/d94ea852-698d-41f1-93e4-df0322a23b47-tuning-conf-dir\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.814083 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/80ac0050-7f55-4845-ace7-a89f2442f884-cni-binary-copy\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.814097 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/d94ea852-698d-41f1-93e4-df0322a23b47-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.814178 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/80ac0050-7f55-4845-ace7-a89f2442f884-multus-daemon-config\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.814949 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/d94ea852-698d-41f1-93e4-df0322a23b47-cni-binary-copy\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.818641 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.828408 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-85hhq\" (UniqueName: \"kubernetes.io/projected/80ac0050-7f55-4845-ace7-a89f2442f884-kube-api-access-85hhq\") pod \"multus-m4dm4\" (UID: \"80ac0050-7f55-4845-ace7-a89f2442f884\") " pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.832350 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxgn8\" (UniqueName: \"kubernetes.io/projected/d94ea852-698d-41f1-93e4-df0322a23b47-kube-api-access-bxgn8\") pod \"multus-additional-cni-plugins-vz5zh\" (UID: \"d94ea852-698d-41f1-93e4-df0322a23b47\") " pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.832356 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.844121 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.858817 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.873702 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.883971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.884045 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.884066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.884094 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.884127 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.887312 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.903819 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.916893 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.939250 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.954063 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.968661 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.980314 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.982308 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-m4dm4" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.986592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.986637 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.986648 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.986664 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.986678 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:01Z","lastTransitionTime":"2025-11-26T16:53:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.989044 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" Nov 26 16:53:01 crc kubenswrapper[4956]: I1126 16:53:01.996360 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:01Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: W1126 16:53:02.001244 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd94ea852_698d_41f1_93e4_df0322a23b47.slice/crio-d1546fb08885e0149e6701643df41c4488ef1dc4656f031463dd0467cf924468 WatchSource:0}: Error finding container d1546fb08885e0149e6701643df41c4488ef1dc4656f031463dd0467cf924468: Status 404 returned error can't find the container with id d1546fb08885e0149e6701643df41c4488ef1dc4656f031463dd0467cf924468 Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.015199 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.030285 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.033637 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f2thc"] Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.035392 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.047796 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.048044 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.048208 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.048323 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.048486 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.048700 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.056979 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.062034 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.083796 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.090591 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.090628 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.090638 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.090656 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.090669 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.103394 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.103405 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.111771 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-vhvtw" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115160 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-systemd\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115206 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-ovn-kubernetes\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115227 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115247 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-bin\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115428 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-netd\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115562 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-kubelet\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115593 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-config\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115667 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46bd4702-b565-4452-be2c-36a3f1f48621-ovn-node-metrics-cert\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.115791 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-etc-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116009 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116153 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-slash\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116179 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-var-lib-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116200 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-ovn\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116257 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-env-overrides\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116284 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-node-log\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116312 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-log-socket\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116335 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-script-lib\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116361 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zzdlc\" (UniqueName: \"kubernetes.io/projected/46bd4702-b565-4452-be2c-36a3f1f48621-kube-api-access-zzdlc\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116385 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-systemd-units\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.116409 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-netns\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.118072 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.133585 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.157435 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.180523 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.180960 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.181005 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.181029 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"35ed81d8c51327c7ee5fe6eca512c96a31989d24cd5773f87ce091bc26f6ef1b"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.182904 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vhvtw" event={"ID":"adc2efe4-3ec6-4e5f-8dd8-76954150acb1","Type":"ContainerStarted","Data":"14b6124bfa07643ae5342c6d2c7c5f48560f15fcf4eb9be94760e1bd80fe599e"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.183944 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerStarted","Data":"d1546fb08885e0149e6701643df41c4488ef1dc4656f031463dd0467cf924468"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.185705 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerStarted","Data":"45ec54fc4529fdee9e0bd64105b49fe54ed8f5fcbe8adea2a32fbf3bf4b22263"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.194317 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.194605 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.194672 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.194684 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.194705 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.194721 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.209493 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.217839 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-systemd\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.217908 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-ovn-kubernetes\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.217939 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.217962 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-netd\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.217992 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-bin\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218015 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-kubelet\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218035 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-config\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218067 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46bd4702-b565-4452-be2c-36a3f1f48621-ovn-node-metrics-cert\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218097 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-etc-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218121 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218158 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-ovn\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218179 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-env-overrides\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218219 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-slash\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218257 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-var-lib-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218282 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-node-log\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218303 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zzdlc\" (UniqueName: \"kubernetes.io/projected/46bd4702-b565-4452-be2c-36a3f1f48621-kube-api-access-zzdlc\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218327 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-log-socket\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218350 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-script-lib\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218384 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-systemd-units\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218406 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-netns\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218485 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-netns\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.218978 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-systemd\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.219021 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-ovn-kubernetes\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.219051 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.219080 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-netd\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.219119 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-bin\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.219148 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-kubelet\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.219949 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-slash\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220063 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-log-socket\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220136 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-systemd-units\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220177 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-var-lib-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220269 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220287 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-node-log\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220354 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-etc-openvswitch\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220396 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-ovn\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.220488 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-config\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.221004 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-env-overrides\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.221406 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-script-lib\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.225448 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.229383 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46bd4702-b565-4452-be2c-36a3f1f48621-ovn-node-metrics-cert\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.238567 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.253570 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.257215 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zzdlc\" (UniqueName: \"kubernetes.io/projected/46bd4702-b565-4452-be2c-36a3f1f48621-kube-api-access-zzdlc\") pod \"ovnkube-node-f2thc\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.269113 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.289027 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.298589 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.298634 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.298649 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.298674 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.298692 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.306606 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.320473 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.338774 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.355060 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.367041 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.368832 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: W1126 16:53:02.380611 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod46bd4702_b565_4452_be2c_36a3f1f48621.slice/crio-7ffa3211be69b02f996dc04517ad911040022ba015618c443c16d8e2b7a4e871 WatchSource:0}: Error finding container 7ffa3211be69b02f996dc04517ad911040022ba015618c443c16d8e2b7a4e871: Status 404 returned error can't find the container with id 7ffa3211be69b02f996dc04517ad911040022ba015618c443c16d8e2b7a4e871 Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.384670 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.403050 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.403113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.403132 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.403161 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.403181 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.403950 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.421279 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.435310 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.461496 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.502179 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.504590 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.504622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.504632 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.504649 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.504686 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.538603 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.607262 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.607307 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.607318 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.607336 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.607350 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.622839 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.622976 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623019 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:53:10.622989418 +0000 UTC m=+36.318949970 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.623073 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623087 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623145 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:10.623132582 +0000 UTC m=+36.319093134 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.623159 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.623186 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623278 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623285 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623283 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623323 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623298 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623330 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:10.623315628 +0000 UTC m=+36.319276230 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623359 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623346 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623391 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:10.623382769 +0000 UTC m=+36.319343321 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.623448 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:10.62340594 +0000 UTC m=+36.319366552 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.709761 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.709807 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.709819 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.709834 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.709846 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.812641 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.812682 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.812696 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.812713 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.812722 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.934658 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.935045 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.935054 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.935072 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.935083 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:02Z","lastTransitionTime":"2025-11-26T16:53:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.999233 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:02 crc kubenswrapper[4956]: E1126 16:53:02.999409 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:02 crc kubenswrapper[4956]: I1126 16:53:02.999923 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:02.999973 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:03 crc kubenswrapper[4956]: E1126 16:53:03.000038 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:03 crc kubenswrapper[4956]: E1126 16:53:03.000140 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.037286 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.037338 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.037353 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.037371 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.037384 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.139912 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.139945 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.139956 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.139974 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.139984 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.192455 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-vhvtw" event={"ID":"adc2efe4-3ec6-4e5f-8dd8-76954150acb1","Type":"ContainerStarted","Data":"8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.194822 4956 generic.go:334] "Generic (PLEG): container finished" podID="d94ea852-698d-41f1-93e4-df0322a23b47" containerID="47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b" exitCode=0 Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.195029 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerDied","Data":"47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.197157 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerStarted","Data":"7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.199168 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" exitCode=0 Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.199220 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.199265 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"7ffa3211be69b02f996dc04517ad911040022ba015618c443c16d8e2b7a4e871"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.211036 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.230497 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.251018 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.254200 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.254257 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.254273 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.254295 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.254315 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.265721 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.288390 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.313281 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.330559 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.348995 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.360054 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.360465 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.360476 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.360498 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.360513 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.374762 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.392847 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.411406 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.424047 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.440356 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.468696 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.471962 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.472078 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.472162 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.472257 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.473841 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.513719 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.538293 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.552295 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.568714 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.577682 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.577726 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.577741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.577761 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.577773 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.591325 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.605691 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.618113 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.643669 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.657747 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.671512 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.680036 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.680078 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.680090 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.680121 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.680132 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.686902 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.699887 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.783764 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.783816 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.783830 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.783879 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.783896 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.886960 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.887001 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.887010 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.887028 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.887039 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.989985 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.990081 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.990125 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.990162 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:03 crc kubenswrapper[4956]: I1126 16:53:03.990186 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:03Z","lastTransitionTime":"2025-11-26T16:53:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.093387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.093479 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.093499 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.093534 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.093553 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.202166 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.202264 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.202304 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.202344 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.202369 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.209893 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.209959 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.209979 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.209998 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.210012 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.210027 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.212548 4956 generic.go:334] "Generic (PLEG): container finished" podID="d94ea852-698d-41f1-93e4-df0322a23b47" containerID="4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709" exitCode=0 Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.212613 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerDied","Data":"4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.236370 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.250412 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.265203 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.281694 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.301571 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.306613 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.306658 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.306675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.306701 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.306721 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.317686 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.332785 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.350153 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.366247 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.382590 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.403927 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.412157 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.412187 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.412199 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.412218 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.412231 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.420315 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.438763 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.515767 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.515821 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.515836 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.515857 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.515897 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.618296 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.618350 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.618366 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.618392 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.618410 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.722633 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.722698 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.722717 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.722741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.722761 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.826597 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.826689 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.826717 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.826754 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.826780 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.929580 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.929664 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.929683 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.929710 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.929728 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:04Z","lastTransitionTime":"2025-11-26T16:53:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.996399 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.996520 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:04 crc kubenswrapper[4956]: E1126 16:53:04.996573 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:04 crc kubenswrapper[4956]: I1126 16:53:04.996644 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:04 crc kubenswrapper[4956]: E1126 16:53:04.996851 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:04 crc kubenswrapper[4956]: E1126 16:53:04.997084 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.021908 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.033589 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.033639 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.033651 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.033671 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.033685 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.042318 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.067922 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.092590 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.113519 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.128363 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.136464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.136522 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.136541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.136570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.136590 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.153528 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.170904 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.201405 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.219285 4956 generic.go:334] "Generic (PLEG): container finished" podID="d94ea852-698d-41f1-93e4-df0322a23b47" containerID="3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d" exitCode=0 Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.219343 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerDied","Data":"3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.225528 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.239316 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.239383 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.239395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.239415 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.239429 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.250602 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.268786 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.291240 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.311774 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.330083 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.342369 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.342392 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.342401 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.342418 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.342430 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.344697 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.358006 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.371297 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.386683 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.400539 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.414097 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.430377 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.481685 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.492354 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.492390 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.492403 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.492428 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.492441 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.500097 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.524040 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.542113 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.596164 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.596202 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.596211 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.596224 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.596259 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.700355 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.700978 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.701064 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.701145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.701225 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.806106 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.806489 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.806642 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.806793 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.806966 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.911639 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.911717 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.911741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.911772 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:05 crc kubenswrapper[4956]: I1126 16:53:05.911792 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:05Z","lastTransitionTime":"2025-11-26T16:53:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.014387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.014434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.014447 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.014469 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.014486 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.119095 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.119159 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.119177 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.119209 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.119231 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.223800 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.223900 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.223928 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.223963 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.223983 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.228371 4956 generic.go:334] "Generic (PLEG): container finished" podID="d94ea852-698d-41f1-93e4-df0322a23b47" containerID="44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da" exitCode=0 Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.228460 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerDied","Data":"44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.238856 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.248110 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.277627 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.296335 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.328115 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.328154 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.328165 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.328183 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.328195 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.329007 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.343168 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.355349 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.372329 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.387953 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.404270 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.419639 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.432460 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.432507 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.432522 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.432541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.432555 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.444755 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.479972 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.503958 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.536020 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.536092 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.536117 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.536148 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.536175 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.640350 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.640433 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.640455 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.640486 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.640507 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.752467 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.752546 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.752574 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.752611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.752638 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.856957 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.857032 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.857056 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.857093 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.857122 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.961117 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.961181 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.961199 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.961225 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.961245 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:06Z","lastTransitionTime":"2025-11-26T16:53:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.994616 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:06 crc kubenswrapper[4956]: E1126 16:53:06.994801 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.994887 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:06 crc kubenswrapper[4956]: E1126 16:53:06.995207 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:06 crc kubenswrapper[4956]: I1126 16:53:06.995571 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:06 crc kubenswrapper[4956]: E1126 16:53:06.995941 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.064529 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.064593 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.064613 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.064640 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.064659 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.168020 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.168091 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.168110 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.168135 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.168154 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.247463 4956 generic.go:334] "Generic (PLEG): container finished" podID="d94ea852-698d-41f1-93e4-df0322a23b47" containerID="948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668" exitCode=0 Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.247519 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerDied","Data":"948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.270099 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.273305 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.273347 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.273364 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.273389 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.273408 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.287801 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.313569 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.346526 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.369273 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.377190 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.377224 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.377241 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.377264 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.377281 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.394049 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.395212 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-vpszs"] Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.395816 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.398694 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.399056 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.399323 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.399789 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.415972 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.438607 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.456579 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.475143 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.480993 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.481041 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.481077 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.481103 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.481119 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.496577 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.509442 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-924q5\" (UniqueName: \"kubernetes.io/projected/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-kube-api-access-924q5\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.509531 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-host\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.509622 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-serviceca\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.516730 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.556795 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.578323 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.584736 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.584796 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.584824 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.584851 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.584894 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.611100 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-924q5\" (UniqueName: \"kubernetes.io/projected/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-kube-api-access-924q5\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.611184 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-host\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.611249 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-serviceca\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.611422 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-host\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.613820 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-serviceca\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.630611 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.655017 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-924q5\" (UniqueName: \"kubernetes.io/projected/8dcb7d83-aba3-4ec1-83f2-2afd638d3911-kube-api-access-924q5\") pod \"node-ca-vpszs\" (UID: \"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\") " pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.660936 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.684795 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.687970 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.688004 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.688014 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.688030 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.688041 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.699456 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.713272 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.715818 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-vpszs" Nov 26 16:53:07 crc kubenswrapper[4956]: W1126 16:53:07.729314 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8dcb7d83_aba3_4ec1_83f2_2afd638d3911.slice/crio-22a305a42657c8edb0a706e9f80d04decf263565eaeab134ea8eb9fb1a090857 WatchSource:0}: Error finding container 22a305a42657c8edb0a706e9f80d04decf263565eaeab134ea8eb9fb1a090857: Status 404 returned error can't find the container with id 22a305a42657c8edb0a706e9f80d04decf263565eaeab134ea8eb9fb1a090857 Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.731018 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.749427 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.760246 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.773426 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.791027 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.791098 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.791114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.791136 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.791150 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.792996 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.808002 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.822340 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.840043 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:07Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.893692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.893734 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.893746 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.893767 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.893783 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.996837 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.996914 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.997152 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.997190 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:07 crc kubenswrapper[4956]: I1126 16:53:07.997204 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:07Z","lastTransitionTime":"2025-11-26T16:53:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.099731 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.099773 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.099786 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.099808 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.099822 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.203228 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.203320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.203339 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.203666 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.203685 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.264423 4956 generic.go:334] "Generic (PLEG): container finished" podID="d94ea852-698d-41f1-93e4-df0322a23b47" containerID="f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba" exitCode=0 Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.264514 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerDied","Data":"f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.268882 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vpszs" event={"ID":"8dcb7d83-aba3-4ec1-83f2-2afd638d3911","Type":"ContainerStarted","Data":"138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.268936 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-vpszs" event={"ID":"8dcb7d83-aba3-4ec1-83f2-2afd638d3911","Type":"ContainerStarted","Data":"22a305a42657c8edb0a706e9f80d04decf263565eaeab134ea8eb9fb1a090857"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.288153 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.309995 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.310043 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.310056 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.310075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.310089 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.310213 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.331261 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.353151 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.374752 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.396492 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.415622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.415707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.415728 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.415755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.415778 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.421773 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.441706 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.465922 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.481826 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.497690 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.516238 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.520523 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.520587 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.520608 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.520637 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.520659 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.536948 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.565389 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.591434 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.610300 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.624620 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.624702 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.624723 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.624757 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.624776 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.628257 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.649463 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.668014 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.682946 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.699711 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.716843 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.727726 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.727812 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.727825 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.727841 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.727854 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.735466 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.755759 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.772505 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.787424 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.813707 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.830950 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.831003 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.831012 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.831029 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.831040 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.841180 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:08Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.935463 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.935551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.935573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.935783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.935803 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:08Z","lastTransitionTime":"2025-11-26T16:53:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.995033 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.995078 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:08 crc kubenswrapper[4956]: I1126 16:53:08.995092 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:08 crc kubenswrapper[4956]: E1126 16:53:08.995247 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:08 crc kubenswrapper[4956]: E1126 16:53:08.995378 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:08 crc kubenswrapper[4956]: E1126 16:53:08.995538 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.039013 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.039057 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.039074 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.039097 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.039117 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.141890 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.141959 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.141982 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.142013 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.142037 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.246103 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.246201 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.246227 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.246265 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.246284 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.279173 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.279735 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.279805 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.288075 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" event={"ID":"d94ea852-698d-41f1-93e4-df0322a23b47","Type":"ContainerStarted","Data":"606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.302024 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.309948 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.317718 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.324616 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.344208 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.350201 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.350279 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.350297 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.350325 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.350343 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.366763 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.391527 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.409583 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.427355 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.448912 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.453634 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.453694 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.453710 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.453737 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.453757 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.477248 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.490943 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.509400 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.528279 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.541767 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.558804 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.558924 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.558948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.558992 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.559027 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.565697 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.581767 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.603626 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.626023 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.641483 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.655423 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.661961 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.662019 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.662038 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.662069 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.662089 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.669245 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.682299 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.696365 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.709899 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.724792 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.746386 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.762394 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.764365 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.764417 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.764440 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.764465 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.764484 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.782051 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.799481 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:09Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.867824 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.867926 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.867948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.867979 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.868001 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.971329 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.971405 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.971424 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.971457 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:09 crc kubenswrapper[4956]: I1126 16:53:09.971478 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:09Z","lastTransitionTime":"2025-11-26T16:53:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.058763 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.075086 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.075160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.075177 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.075200 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.075216 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.079975 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.105626 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.129529 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.147279 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.171782 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.178692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.178779 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.178798 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.178828 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.178849 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.193814 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.220592 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.248046 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.271400 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.282748 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.282805 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.282824 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.282850 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.282909 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.292763 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.293478 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.316465 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.337755 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.352933 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.373018 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:10Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.386705 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.386781 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.386803 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.386831 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.386852 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.490318 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.490401 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.490434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.490471 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.490504 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.594215 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.594266 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.594275 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.594293 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.594302 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.645822 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.646061 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646111 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:53:26.646070588 +0000 UTC m=+52.342031180 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.646168 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646194 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646293 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:26.646264443 +0000 UTC m=+52.342225035 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.646331 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.646382 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646485 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646539 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646556 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646562 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646583 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646606 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646636 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646659 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:26.646643334 +0000 UTC m=+52.342603916 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646781 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:26.646760487 +0000 UTC m=+52.342721069 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.646821 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:26.646807288 +0000 UTC m=+52.342767880 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.698145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.698220 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.698241 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.698271 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.698291 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.801732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.801795 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.801818 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.801851 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.801908 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.905335 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.905382 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.905395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.905416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.905428 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:10Z","lastTransitionTime":"2025-11-26T16:53:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.995399 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.995464 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:10 crc kubenswrapper[4956]: I1126 16:53:10.995475 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.995577 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.995765 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:10 crc kubenswrapper[4956]: E1126 16:53:10.996105 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.007411 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.007457 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.007474 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.007492 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.007507 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.110023 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.110063 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.110079 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.110100 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.110114 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.213315 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.213391 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.213412 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.213457 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.213482 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.296142 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.317012 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.317078 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.317097 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.317127 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.317146 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.369616 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.369685 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.369705 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.369734 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.369755 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: E1126 16:53:11.392281 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:11Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.398195 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.398252 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.398270 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.398301 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.398319 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: E1126 16:53:11.415710 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:11Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.426353 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.426419 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.426450 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.426561 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.426717 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: E1126 16:53:11.453496 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:11Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.458765 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.458827 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.458846 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.458915 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.458940 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: E1126 16:53:11.478530 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:11Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.484313 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.484366 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.484391 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.484454 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.484477 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: E1126 16:53:11.501202 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:11Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:11 crc kubenswrapper[4956]: E1126 16:53:11.501461 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.503557 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.503613 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.503622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.503641 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.503654 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.606964 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.607038 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.607057 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.607089 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.607111 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.711192 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.711570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.711592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.711621 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.711639 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.814630 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.814916 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.814947 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.814984 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.815008 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.918776 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.918848 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.918903 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.918937 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:11 crc kubenswrapper[4956]: I1126 16:53:11.918956 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:11Z","lastTransitionTime":"2025-11-26T16:53:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.022321 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.022391 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.022411 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.022443 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.022465 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.125921 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.125975 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.125991 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.126021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.126043 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.229431 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.229517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.229539 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.229567 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.229585 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.303354 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/0.log" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.308133 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d" exitCode=1 Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.308232 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.309706 4956 scope.go:117] "RemoveContainer" containerID="2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.336371 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.336458 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.336483 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.336518 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.336545 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.338024 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.360259 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.383018 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.409406 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.432833 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.441655 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.441700 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.441714 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.441735 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.441749 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.453355 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.475530 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.493798 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.520175 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.545274 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.545312 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.545324 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.545347 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.545359 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.548893 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.566373 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.587302 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.604780 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.621978 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:12Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.649450 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.649547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.649574 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.649628 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.649655 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.752434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.752466 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.752474 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.752487 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.752495 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.854811 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.854876 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.854890 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.854908 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.854921 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.958037 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.958087 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.958105 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.958133 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.958150 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:12Z","lastTransitionTime":"2025-11-26T16:53:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.995170 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.995232 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:12 crc kubenswrapper[4956]: E1126 16:53:12.995311 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:12 crc kubenswrapper[4956]: I1126 16:53:12.995363 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:12 crc kubenswrapper[4956]: E1126 16:53:12.995577 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:12 crc kubenswrapper[4956]: E1126 16:53:12.995757 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.060370 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.060419 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.060431 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.060450 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.060462 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.163021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.163080 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.163092 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.163111 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.163125 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.266337 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.266387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.266400 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.266419 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.266430 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.313020 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/0.log" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.315951 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.316110 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.331289 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.346135 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.359237 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.368459 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.368495 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.368507 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.368527 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.368540 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.372560 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.384583 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.397999 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.408273 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.420547 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.431178 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.440418 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.453927 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.471257 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.471295 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.471307 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.471326 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.471339 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.472285 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.483656 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.503681 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:13Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.574397 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.574454 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.574465 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.574487 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.574499 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.676819 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.676882 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.676892 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.676906 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.676919 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.780151 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.780210 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.780222 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.780240 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.780252 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.883354 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.883416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.883428 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.883450 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.883464 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.991907 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.991966 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.991984 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.992010 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:13 crc kubenswrapper[4956]: I1126 16:53:13.992034 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:13Z","lastTransitionTime":"2025-11-26T16:53:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.095160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.095467 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.095552 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.095648 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.095786 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.198302 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.198358 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.198368 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.198388 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.198402 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.302312 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.302363 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.302372 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.302387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.302398 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.321718 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/1.log" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.322681 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/0.log" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.326157 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8" exitCode=1 Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.326228 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.326288 4956 scope.go:117] "RemoveContainer" containerID="2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.329655 4956 scope.go:117] "RemoveContainer" containerID="bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8" Nov 26 16:53:14 crc kubenswrapper[4956]: E1126 16:53:14.330101 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.348472 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.366854 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.384047 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr"] Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.386058 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.389266 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.389490 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.396328 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.406614 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.406669 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.406688 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.406714 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.406733 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.429574 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.449136 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.470719 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.494445 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.494508 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.494700 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7zg7\" (UniqueName: \"kubernetes.io/projected/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-kube-api-access-k7zg7\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.494752 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.496519 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.510387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.510478 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.510499 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.510529 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.510548 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.517155 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.540257 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.560306 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.581983 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.595717 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7zg7\" (UniqueName: \"kubernetes.io/projected/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-kube-api-access-k7zg7\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.595807 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.595909 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.595945 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.597008 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-env-overrides\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.597089 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.601385 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.605950 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.614690 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.614734 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.614751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.614780 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.614797 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.628091 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7zg7\" (UniqueName: \"kubernetes.io/projected/8bfae4c3-3d62-4968-91d4-8a8caabc4af8-kube-api-access-k7zg7\") pod \"ovnkube-control-plane-749d76644c-f6knr\" (UID: \"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.631229 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.650965 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.667896 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.684993 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.701300 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.708167 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.718114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.718167 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.718188 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.718213 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.718233 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.725913 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: W1126 16:53:14.733073 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bfae4c3_3d62_4968_91d4_8a8caabc4af8.slice/crio-e4690546b49667164dc2dc2d342b8ac0d9906b7eebf96a75a479c6a726ba4128 WatchSource:0}: Error finding container e4690546b49667164dc2dc2d342b8ac0d9906b7eebf96a75a479c6a726ba4128: Status 404 returned error can't find the container with id e4690546b49667164dc2dc2d342b8ac0d9906b7eebf96a75a479c6a726ba4128 Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.756391 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.776829 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.802363 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.821416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.821478 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.821493 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.821520 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.821535 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.829725 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.854010 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.875139 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.891646 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.913699 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.925212 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.925270 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.925282 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.925302 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.925315 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:14Z","lastTransitionTime":"2025-11-26T16:53:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.930860 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.948242 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.965609 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:14Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.995591 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:14 crc kubenswrapper[4956]: E1126 16:53:14.995792 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.996334 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:14 crc kubenswrapper[4956]: E1126 16:53:14.996445 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:14 crc kubenswrapper[4956]: I1126 16:53:14.996590 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:14 crc kubenswrapper[4956]: E1126 16:53:14.996696 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.022409 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.028701 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.028922 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.028994 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.029018 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.029035 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.045301 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.073250 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.108904 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.133111 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.133164 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.133183 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.133212 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.133232 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.133482 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.157994 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.183974 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.204283 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.236711 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.236793 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.236817 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.236846 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.236886 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.239031 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.262476 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.285694 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.312594 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.332582 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.333658 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/1.log" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339056 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339066 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" event={"ID":"8bfae4c3-3d62-4968-91d4-8a8caabc4af8","Type":"ContainerStarted","Data":"884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339300 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339365 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" event={"ID":"8bfae4c3-3d62-4968-91d4-8a8caabc4af8","Type":"ContainerStarted","Data":"27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339405 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" event={"ID":"8bfae4c3-3d62-4968-91d4-8a8caabc4af8","Type":"ContainerStarted","Data":"e4690546b49667164dc2dc2d342b8ac0d9906b7eebf96a75a479c6a726ba4128"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339764 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.339782 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.353526 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.370419 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.388960 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.404159 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.420534 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.437204 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.441912 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.441951 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.441961 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.441978 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.441990 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.451540 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.466956 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.486093 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.510575 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.518284 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-b8nt2"] Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.518780 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:15 crc kubenswrapper[4956]: E1126 16:53:15.518837 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.527426 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.541911 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.544887 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.544949 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.544971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.545003 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.545024 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.561814 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.578182 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.602567 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.608760 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.609072 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nc92\" (UniqueName: \"kubernetes.io/projected/0956e12e-96b8-4b83-bc0a-666639b36586-kube-api-access-6nc92\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.633003 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.648661 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.648723 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.648745 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.648773 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.648805 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.656580 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.674501 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.692853 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.708086 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.710691 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.710931 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nc92\" (UniqueName: \"kubernetes.io/projected/0956e12e-96b8-4b83-bc0a-666639b36586-kube-api-access-6nc92\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:15 crc kubenswrapper[4956]: E1126 16:53:15.710966 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:15 crc kubenswrapper[4956]: E1126 16:53:15.711080 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:16.211043941 +0000 UTC m=+41.907004533 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.730358 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nc92\" (UniqueName: \"kubernetes.io/projected/0956e12e-96b8-4b83-bc0a-666639b36586-kube-api-access-6nc92\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.737670 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.754164 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.754237 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.754263 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.754297 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.754322 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.759605 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.785307 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.810348 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.845146 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2c182f34e030388c60b32dddb56a9c4c79eb34f95056826345ea580420f01d9d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:11Z\\\",\\\"message\\\":\\\"ed *v1.Pod event handler 6\\\\nI1126 16:53:11.598021 6284 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1126 16:53:11.598339 6284 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:53:11.598804 6284 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI1126 16:53:11.598846 6284 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI1126 16:53:11.598916 6284 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:11.598919 6284 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI1126 16:53:11.598947 6284 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI1126 16:53:11.598957 6284 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1126 16:53:11.599004 6284 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:11.599059 6284 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:11.599071 6284 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:11.599129 6284 factory.go:656] Stopping watch factory\\\\nI1126 16:53:11.599134 6284 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI1126 16:53:11.599154 6284 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:11.599156 6284 handler.go:208] Removed *v1.Node event handler 2\\\\nI11\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.857783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.857857 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.857915 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.857954 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.857980 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.861842 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.881509 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.904793 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.929422 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.952650 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.962127 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.962187 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.962216 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.962252 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.962279 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:15Z","lastTransitionTime":"2025-11-26T16:53:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.976196 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:15 crc kubenswrapper[4956]: I1126 16:53:15.999662 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:15Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.021546 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:16Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.065497 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.065546 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.065566 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.065592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.065616 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.169211 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.169287 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.169306 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.169336 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.169361 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.218145 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:16 crc kubenswrapper[4956]: E1126 16:53:16.218448 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:16 crc kubenswrapper[4956]: E1126 16:53:16.218576 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:17.218543177 +0000 UTC m=+42.914503769 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.272543 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.272612 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.272637 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.272669 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.272694 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.375519 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.375599 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.375622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.375647 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.375665 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.478805 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.478902 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.478923 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.478950 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.478970 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.581739 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.581813 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.581833 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.581860 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.581907 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.684628 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.684687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.684706 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.684732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.684750 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.787797 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.787941 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.787963 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.787993 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.788015 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.891262 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.891321 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.891335 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.891356 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.891373 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994273 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994355 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:16 crc kubenswrapper[4956]: E1126 16:53:16.994388 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994264 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994428 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994453 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994502 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994502 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:16 crc kubenswrapper[4956]: I1126 16:53:16.994520 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:16Z","lastTransitionTime":"2025-11-26T16:53:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:16 crc kubenswrapper[4956]: E1126 16:53:16.994678 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:16 crc kubenswrapper[4956]: E1126 16:53:16.994734 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:16 crc kubenswrapper[4956]: E1126 16:53:16.994796 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.100153 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.100367 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.100395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.100471 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.100504 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.204330 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.204391 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.204410 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.204433 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.204453 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.231804 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:17 crc kubenswrapper[4956]: E1126 16:53:17.232025 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:17 crc kubenswrapper[4956]: E1126 16:53:17.232117 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:19.232094137 +0000 UTC m=+44.928054699 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.307333 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.307395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.307414 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.307441 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.307459 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.411568 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.411628 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.411645 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.411669 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.411690 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.515267 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.515383 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.515396 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.515417 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.515433 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.619035 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.619124 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.619149 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.619183 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.619207 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.722540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.722608 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.722625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.722651 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.722671 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.826464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.826533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.826555 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.826580 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.826597 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.930547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.930635 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.930657 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.930686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:17 crc kubenswrapper[4956]: I1126 16:53:17.930705 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:17Z","lastTransitionTime":"2025-11-26T16:53:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.033626 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.033688 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.033701 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.033720 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.033736 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.136980 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.137047 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.137059 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.137080 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.137091 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.240196 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.240268 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.240288 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.240320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.240342 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.343499 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.343595 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.343622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.343650 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.343673 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.446832 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.446926 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.446945 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.446971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.446999 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.549648 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.549732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.549755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.549783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.549813 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.653617 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.653681 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.653693 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.653710 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.653722 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.756777 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.756843 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.756889 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.756921 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.756946 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.859848 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.859935 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.859953 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.859976 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.859997 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.962546 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.962624 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.962648 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.962676 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.962696 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:18Z","lastTransitionTime":"2025-11-26T16:53:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.994382 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.994448 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.994532 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:18 crc kubenswrapper[4956]: I1126 16:53:18.994765 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:18 crc kubenswrapper[4956]: E1126 16:53:18.994746 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:18 crc kubenswrapper[4956]: E1126 16:53:18.994886 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:18 crc kubenswrapper[4956]: E1126 16:53:18.995098 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:18 crc kubenswrapper[4956]: E1126 16:53:18.995339 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.066839 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.066949 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.066968 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.067024 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.067043 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.170428 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.170499 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.170519 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.170547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.170569 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.255704 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:19 crc kubenswrapper[4956]: E1126 16:53:19.255984 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:19 crc kubenswrapper[4956]: E1126 16:53:19.256100 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:23.256071031 +0000 UTC m=+48.952031623 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.273430 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.273500 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.273517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.273542 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.273564 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.376438 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.376515 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.376533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.376562 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.376588 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.479822 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.479924 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.479955 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.479987 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.480007 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.583233 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.583307 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.583326 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.583351 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.583369 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.687604 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.687675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.687693 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.687724 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.687743 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.790420 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.790492 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.790510 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.790536 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.790559 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.893740 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.893812 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.893833 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.893893 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.893918 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.998023 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.998082 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.998099 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.998126 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:19 crc kubenswrapper[4956]: I1126 16:53:19.998147 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:19Z","lastTransitionTime":"2025-11-26T16:53:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.101295 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.101361 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.101381 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.101408 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.101429 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.206304 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.206380 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.206398 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.206426 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.206448 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.310363 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.310444 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.310472 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.310500 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.310518 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.413970 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.414041 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.414061 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.414089 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.414108 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.517461 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.517539 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.517557 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.517586 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.517608 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.621139 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.621220 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.621237 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.621265 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.621283 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.725699 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.725783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.725808 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.725839 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.725890 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.829755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.829823 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.829841 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.829904 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.829924 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.932509 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.932564 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.932582 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.932605 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.932621 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:20Z","lastTransitionTime":"2025-11-26T16:53:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.995239 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.995275 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.995305 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:20 crc kubenswrapper[4956]: E1126 16:53:20.995471 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:20 crc kubenswrapper[4956]: I1126 16:53:20.995486 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:20 crc kubenswrapper[4956]: E1126 16:53:20.995605 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:20 crc kubenswrapper[4956]: E1126 16:53:20.995721 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:20 crc kubenswrapper[4956]: E1126 16:53:20.995805 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.037227 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.037337 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.037359 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.037434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.037459 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.141334 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.141420 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.141454 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.141486 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.141507 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.244663 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.244736 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.244753 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.244777 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.244795 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.348712 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.348817 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.348841 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.348917 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.348954 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.452290 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.452362 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.452390 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.452424 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.452449 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.517517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.517590 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.517606 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.517632 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.517651 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: E1126 16:53:21.538749 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:21Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.544442 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.544505 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.544525 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.544555 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.544575 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: E1126 16:53:21.567402 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:21Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.574808 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.575100 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.575130 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.575160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.575218 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: E1126 16:53:21.598681 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:21Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.604903 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.604973 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.604991 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.605022 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.605041 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: E1126 16:53:21.626415 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:21Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.631830 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.631923 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.631936 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.631954 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.631995 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: E1126 16:53:21.660656 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:21Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:21 crc kubenswrapper[4956]: E1126 16:53:21.664627 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.672195 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.672277 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.672299 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.672326 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.672344 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.775948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.776018 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.776040 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.776076 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.776100 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.879924 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.880004 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.880025 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.880055 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.880074 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.983417 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.983589 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.983611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.983638 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:21 crc kubenswrapper[4956]: I1126 16:53:21.983656 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:21Z","lastTransitionTime":"2025-11-26T16:53:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.086994 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.087448 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.087690 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.087922 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.088082 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.192455 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.192530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.192547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.192577 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.192598 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.262433 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.263970 4956 scope.go:117] "RemoveContainer" containerID="bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8" Nov 26 16:53:22 crc kubenswrapper[4956]: E1126 16:53:22.264259 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.288254 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.297145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.297201 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.297219 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.297249 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.297269 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.311664 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.329931 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.355194 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.388293 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.400986 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.401054 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.401073 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.401098 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.401116 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.404574 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.424205 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.445729 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.465231 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.485061 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.504232 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.504719 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.504774 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.504792 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.504820 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.504838 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.524446 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.546220 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.565809 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.583366 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.597781 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:22Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.608771 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.608831 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.608849 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.608915 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.608938 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.712088 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.712163 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.712188 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.712221 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.712244 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.816196 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.816266 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.816285 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.816314 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.816334 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.919579 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.919655 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.919679 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.919713 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.919738 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:22Z","lastTransitionTime":"2025-11-26T16:53:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.995414 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.995572 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.995446 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:22 crc kubenswrapper[4956]: I1126 16:53:22.995658 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:22 crc kubenswrapper[4956]: E1126 16:53:22.995654 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:22 crc kubenswrapper[4956]: E1126 16:53:22.995792 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:22 crc kubenswrapper[4956]: E1126 16:53:22.995918 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:22 crc kubenswrapper[4956]: E1126 16:53:22.996076 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.023335 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.023398 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.023420 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.023448 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.023469 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.126147 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.126215 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.126231 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.126253 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.126266 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.230684 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.230758 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.230775 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.230800 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.230819 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.308006 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:23 crc kubenswrapper[4956]: E1126 16:53:23.308307 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:23 crc kubenswrapper[4956]: E1126 16:53:23.308436 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:31.308402159 +0000 UTC m=+57.004362751 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.334189 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.334248 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.334265 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.334289 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.334308 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.438116 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.438185 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.438206 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.438248 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.438271 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.541632 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.541692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.541704 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.541723 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.541736 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.646533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.646624 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.646646 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.646764 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.646793 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.750549 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.750657 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.750678 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.750707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.750731 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.854685 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.854763 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.854787 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.854817 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.854838 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.959174 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.959263 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.959283 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.959316 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:23 crc kubenswrapper[4956]: I1126 16:53:23.959337 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:23Z","lastTransitionTime":"2025-11-26T16:53:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.063611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.063688 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.063707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.063736 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.063756 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.168162 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.168246 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.168267 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.168298 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.168319 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.272016 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.272083 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.272103 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.272139 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.272167 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.375569 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.375629 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.375648 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.375677 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.375695 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.479281 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.479369 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.479387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.479427 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.479448 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.582843 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.582976 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.582994 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.583020 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.583042 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.686624 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.686703 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.686721 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.686750 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.686772 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.790613 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.790687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.790707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.790735 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.790754 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.894366 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.894433 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.894451 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.894478 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.894498 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.994541 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.994575 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.994629 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:24 crc kubenswrapper[4956]: E1126 16:53:24.994814 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.994931 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:24 crc kubenswrapper[4956]: E1126 16:53:24.995165 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:24 crc kubenswrapper[4956]: E1126 16:53:24.995441 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:24 crc kubenswrapper[4956]: E1126 16:53:24.995665 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.998476 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.998534 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.998556 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.998586 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:24 crc kubenswrapper[4956]: I1126 16:53:24.998611 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:24Z","lastTransitionTime":"2025-11-26T16:53:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.019822 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.039395 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.060708 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.083245 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.098764 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.101846 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.101980 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.102020 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.102050 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.102069 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.117053 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.137773 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.154433 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.173201 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.188994 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.203817 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.205746 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.205795 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.205813 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.205834 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.205850 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.224850 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.246579 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.263859 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.290629 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.308991 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.309086 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.309104 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.309156 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.309176 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.316121 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:25Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.411305 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.411359 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.411371 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.411391 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.411404 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.514651 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.514699 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.514709 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.514723 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.514734 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.617202 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.617257 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.617265 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.617280 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.617289 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.719759 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.719803 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.719814 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.719832 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.719843 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.823439 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.823511 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.823529 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.823982 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.824275 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.928113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.928179 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.928199 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.928225 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:25 crc kubenswrapper[4956]: I1126 16:53:25.928242 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:25Z","lastTransitionTime":"2025-11-26T16:53:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.031465 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.031512 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.031529 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.031551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.031567 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.134858 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.134932 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.134944 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.134961 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.134973 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.238988 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.239056 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.239074 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.239101 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.239121 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.342816 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.342942 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.342971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.342997 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.343015 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.445934 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.446005 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.446026 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.446054 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.446073 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.549086 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.549160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.549180 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.549216 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.549241 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.653111 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.653165 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.653186 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.653212 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.653237 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.744677 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.745358 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.745481 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.745538 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.745634 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747348 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747510 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747578 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747608 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747522 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:58.747482543 +0000 UTC m=+84.443443125 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747517 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747754 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:53:58.747706029 +0000 UTC m=+84.443666611 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747794 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747808 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:58.747790831 +0000 UTC m=+84.443751413 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747836 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747849 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:58.747830562 +0000 UTC m=+84.443791144 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.747901 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.748044 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:58.747999067 +0000 UTC m=+84.443959649 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.756790 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.756885 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.756905 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.756932 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.756950 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.860315 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.860422 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.860441 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.860467 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.860490 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.964127 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.964203 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.964221 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.964250 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.964267 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:26Z","lastTransitionTime":"2025-11-26T16:53:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.995153 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.995719 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.995469 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.995861 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.995512 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.995996 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:26 crc kubenswrapper[4956]: I1126 16:53:26.995361 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:26 crc kubenswrapper[4956]: E1126 16:53:26.996093 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.068250 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.068626 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.068750 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.068916 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.069044 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.173208 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.173289 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.173307 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.173334 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.173354 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.276721 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.276799 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.276820 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.276847 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.276904 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.381389 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.381471 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.381490 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.381518 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.381538 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.485273 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.485618 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.485743 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.485906 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.486130 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.589449 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.589543 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.589570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.589603 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.589625 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.691636 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.691684 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.691696 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.691712 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.691724 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.795573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.795633 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.795655 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.795681 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.795700 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.898453 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.898517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.898533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.898613 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:27 crc kubenswrapper[4956]: I1126 16:53:27.898630 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:27Z","lastTransitionTime":"2025-11-26T16:53:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.002162 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.002238 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.002259 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.002287 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.002307 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.105895 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.105959 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.105970 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.105990 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.105999 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.209103 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.209175 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.209193 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.209217 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.209235 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.311618 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.311652 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.311661 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.311692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.311702 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.415529 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.415600 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.415625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.416066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.416113 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.519149 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.519224 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.519241 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.519269 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.519288 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.623118 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.623186 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.623205 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.623229 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.623246 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.726198 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.726299 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.726317 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.726341 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.726360 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.831583 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.831653 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.831669 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.831697 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.831713 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.935447 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.935517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.935541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.935573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.935598 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:28Z","lastTransitionTime":"2025-11-26T16:53:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.994829 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.994861 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:28 crc kubenswrapper[4956]: E1126 16:53:28.995031 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.995092 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:28 crc kubenswrapper[4956]: I1126 16:53:28.995134 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:28 crc kubenswrapper[4956]: E1126 16:53:28.995270 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:28 crc kubenswrapper[4956]: E1126 16:53:28.995472 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:28 crc kubenswrapper[4956]: E1126 16:53:28.995678 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.039029 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.039095 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.039113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.039137 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.039157 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.143110 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.143178 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.143197 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.143226 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.143246 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.247043 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.247104 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.247114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.247131 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.247143 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.350359 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.350407 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.350418 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.350434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.350446 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.453650 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.453741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.453760 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.453785 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.453801 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.558829 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.558929 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.558952 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.558985 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.559011 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.662039 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.662094 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.662106 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.662129 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.662144 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.765067 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.765125 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.765140 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.765160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.765175 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.863729 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.867818 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.867886 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.867899 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.867919 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.867937 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.884569 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.894910 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:29Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.923471 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:29Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.943186 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:29Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.968025 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:29Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.970927 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.970978 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.970995 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.971025 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.971045 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:29Z","lastTransitionTime":"2025-11-26T16:53:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:29 crc kubenswrapper[4956]: I1126 16:53:29.991748 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:29Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.011160 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.032529 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.048337 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.073287 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.073325 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.073336 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.073355 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.073368 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.074218 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.088406 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.130072 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.156266 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.176193 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.176247 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.176266 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.176291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.176308 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.187899 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.204499 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.233820 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.247257 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:30Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.279409 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.279491 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.279515 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.279547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.279569 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.383187 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.383282 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.383307 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.383344 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.383370 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.486464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.486531 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.486555 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.486588 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.486611 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.589414 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.589448 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.589459 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.589475 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.589487 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.695347 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.695416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.695446 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.695481 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.695507 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.799066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.799130 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.799148 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.799174 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.799193 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.903096 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.903186 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.903206 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.903235 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.903254 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:30Z","lastTransitionTime":"2025-11-26T16:53:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.994657 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.994684 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.994773 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:30 crc kubenswrapper[4956]: I1126 16:53:30.994833 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:30 crc kubenswrapper[4956]: E1126 16:53:30.995126 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:30 crc kubenswrapper[4956]: E1126 16:53:30.995225 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:30 crc kubenswrapper[4956]: E1126 16:53:30.995322 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:30 crc kubenswrapper[4956]: E1126 16:53:30.995512 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.006630 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.006724 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.006783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.006810 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.006861 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.109965 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.110031 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.110050 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.110079 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.110100 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.213519 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.213581 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.213602 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.213629 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.213650 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.314788 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:31 crc kubenswrapper[4956]: E1126 16:53:31.315101 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:31 crc kubenswrapper[4956]: E1126 16:53:31.315210 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:53:47.315181655 +0000 UTC m=+73.011142237 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.317278 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.317334 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.317352 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.317377 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.317394 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.420822 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.420941 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.420969 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.421000 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.421022 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.524030 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.524113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.524137 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.524170 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.524201 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.627222 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.627276 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.627293 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.627318 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.627336 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.730139 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.730221 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.730283 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.730320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.730347 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.833426 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.833490 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.833507 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.833533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.833552 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.874381 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.874432 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.874446 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.874469 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.874485 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: E1126 16:53:31.899696 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:31Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.906144 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.906227 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.906247 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.906279 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.906297 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: E1126 16:53:31.928403 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:31Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.933703 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.933761 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.933772 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.933792 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.933809 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: E1126 16:53:31.954557 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:31Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.959350 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.959447 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.959464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.959514 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.959533 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:31 crc kubenswrapper[4956]: E1126 16:53:31.981768 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:31Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.986957 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.987019 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.987040 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.987073 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:31 crc kubenswrapper[4956]: I1126 16:53:31.987097 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:31Z","lastTransitionTime":"2025-11-26T16:53:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: E1126 16:53:32.010496 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:32Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:32 crc kubenswrapper[4956]: E1126 16:53:32.010656 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.012687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.012755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.012774 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.012802 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.012822 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.116464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.116517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.116526 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.116543 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.116556 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.219776 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.219853 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.219898 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.219925 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.219946 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.323430 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.323501 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.323520 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.323547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.323565 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.426723 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.426791 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.426809 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.426836 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.426850 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.530219 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.530304 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.530321 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.530344 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.530359 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.633899 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.633989 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.634012 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.634047 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.634071 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.736497 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.736536 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.736544 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.736556 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.736566 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.839278 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.839344 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.839360 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.839386 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.839402 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.942199 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.942263 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.942282 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.942311 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.942332 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:32Z","lastTransitionTime":"2025-11-26T16:53:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.995153 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.995215 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.995172 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:32 crc kubenswrapper[4956]: E1126 16:53:32.995435 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:32 crc kubenswrapper[4956]: I1126 16:53:32.995495 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:32 crc kubenswrapper[4956]: E1126 16:53:32.995636 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:32 crc kubenswrapper[4956]: E1126 16:53:32.995769 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:32 crc kubenswrapper[4956]: E1126 16:53:32.995890 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.045468 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.045540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.045557 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.045587 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.045607 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.148847 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.149047 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.149069 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.149099 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.149120 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.252356 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.252442 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.252463 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.252493 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.252513 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.355681 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.355741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.355753 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.355775 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.355788 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.459037 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.459095 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.459110 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.459133 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.459150 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.562320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.562387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.562408 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.562436 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.562457 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.666475 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.666551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.666567 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.666595 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.666615 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.770524 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.770592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.770610 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.770643 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.770666 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.873814 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.873939 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.873965 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.874004 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.874025 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.978493 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.978581 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.978602 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.978641 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:33 crc kubenswrapper[4956]: I1126 16:53:33.978662 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:33Z","lastTransitionTime":"2025-11-26T16:53:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.082951 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.083449 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.083491 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.083530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.083554 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.187272 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.187338 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.187354 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.187380 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.187399 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.291105 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.291182 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.291202 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.291235 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.291260 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.394673 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.394734 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.394755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.394778 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.394797 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.498781 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.498849 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.498897 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.498928 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.498951 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.602087 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.602141 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.602153 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.602172 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.602188 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.705308 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.705361 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.705373 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.705395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.705408 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.809064 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.809147 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.809174 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.809212 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.809241 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.912394 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.912469 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.912490 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.912526 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.912548 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:34Z","lastTransitionTime":"2025-11-26T16:53:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.994711 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.994860 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.995070 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:34 crc kubenswrapper[4956]: E1126 16:53:34.995022 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:34 crc kubenswrapper[4956]: E1126 16:53:34.995453 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:34 crc kubenswrapper[4956]: I1126 16:53:34.995541 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:34 crc kubenswrapper[4956]: E1126 16:53:34.995770 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:34 crc kubenswrapper[4956]: E1126 16:53:34.995937 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.015558 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.015655 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.015684 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.015725 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.015754 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.025475 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.051083 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.077698 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.102982 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.119222 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.119292 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.119317 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.119358 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.119383 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.129122 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.153183 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.175972 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.196834 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.214440 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.223366 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.223440 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.223467 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.223505 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.223532 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.235023 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.252570 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.277489 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.295070 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.310929 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.327107 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.327373 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.327458 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.327553 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.327651 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.339532 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.375312 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.391129 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:35Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.430608 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.431058 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.431261 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.431435 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.431603 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.535374 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.535466 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.535490 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.535522 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.535549 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.638852 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.638998 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.639033 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.639074 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.639102 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.743196 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.743333 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.743356 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.743386 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.743407 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.846937 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.847015 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.847033 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.847058 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.847071 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.949998 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.950075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.950089 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.950113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.950129 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:35Z","lastTransitionTime":"2025-11-26T16:53:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:35 crc kubenswrapper[4956]: I1126 16:53:35.996666 4956 scope.go:117] "RemoveContainer" containerID="bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.054421 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.055309 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.055491 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.055640 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.055797 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.160003 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.160087 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.160110 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.160145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.160170 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.263569 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.263623 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.263643 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.263670 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.263689 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.371289 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.371346 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.371368 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.371390 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.371407 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.430960 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/1.log" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.433624 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.434180 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.451854 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.471789 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.474626 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.474673 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.474687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.474708 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.474723 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.486807 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.503753 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.520263 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.540940 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.559585 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.578524 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.578586 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.578607 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.578637 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.578659 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.579508 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.598801 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.614453 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.643793 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.674836 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.681035 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.681083 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.681099 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.681118 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.681161 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.693413 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.720762 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.740275 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.753559 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.770578 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:36Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.784151 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.784214 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.784223 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.784242 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.784254 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.886981 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.887044 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.887061 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.887088 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.887107 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.989994 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.990062 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.990075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.990104 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.990121 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:36Z","lastTransitionTime":"2025-11-26T16:53:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.994326 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.994351 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.994325 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:36 crc kubenswrapper[4956]: I1126 16:53:36.994324 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:36 crc kubenswrapper[4956]: E1126 16:53:36.994505 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:36 crc kubenswrapper[4956]: E1126 16:53:36.994648 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:36 crc kubenswrapper[4956]: E1126 16:53:36.994746 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:36 crc kubenswrapper[4956]: E1126 16:53:36.994901 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.093673 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.093736 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.093749 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.093769 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.093785 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.197352 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.197424 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.197444 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.197476 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.197496 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.301223 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.301279 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.301291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.301317 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.301334 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.404345 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.404393 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.404404 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.404423 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.404434 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.507641 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.507733 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.507753 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.507780 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.507800 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.611172 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.611230 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.611241 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.611268 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.611280 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.713968 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.714022 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.714066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.714085 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.714098 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.818058 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.818136 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.818154 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.818185 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.818205 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.921101 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.921156 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.921170 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.921187 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:37 crc kubenswrapper[4956]: I1126 16:53:37.921198 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:37Z","lastTransitionTime":"2025-11-26T16:53:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.025290 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.025348 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.025363 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.025385 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.025400 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.128572 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.128640 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.128659 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.128687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.128708 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.232575 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.232650 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.232666 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.232691 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.232707 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.337278 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.337338 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.337353 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.337375 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.337391 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.440675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.440793 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.440904 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.440938 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.440959 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.444364 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/2.log" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.445449 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/1.log" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.449703 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b" exitCode=1 Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.449780 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.449837 4956 scope.go:117] "RemoveContainer" containerID="bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.451204 4956 scope.go:117] "RemoveContainer" containerID="72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b" Nov 26 16:53:38 crc kubenswrapper[4956]: E1126 16:53:38.451470 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.476495 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.506922 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.528217 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.545521 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.545570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.545588 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.545611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.545629 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.552057 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.572680 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.591190 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.612083 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.629254 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.648650 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.653070 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.653137 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.653159 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.653607 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.653666 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.669772 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.687552 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.714387 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.735632 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.753070 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.760520 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.760606 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.760624 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.760654 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.760676 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.779644 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.809444 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.828859 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:38Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.865528 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.865666 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.865722 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.865751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.865803 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.970140 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.970211 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.970237 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.970271 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.970293 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:38Z","lastTransitionTime":"2025-11-26T16:53:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:38 crc kubenswrapper[4956]: I1126 16:53:38.998824 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:38 crc kubenswrapper[4956]: E1126 16:53:38.999500 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.004431 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.004486 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:39 crc kubenswrapper[4956]: E1126 16:53:39.004626 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.004689 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:39 crc kubenswrapper[4956]: E1126 16:53:39.004906 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:39 crc kubenswrapper[4956]: E1126 16:53:39.005084 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.073304 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.073377 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.073396 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.073424 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.073445 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.176884 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.176937 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.176948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.176969 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.176984 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.280103 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.280570 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.280593 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.280624 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.280646 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.384235 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.384283 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.384295 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.384315 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.384331 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.455981 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/2.log" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.485914 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.485992 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.486021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.486055 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.486083 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.588678 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.588738 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.588755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.588789 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.588810 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.691711 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.691790 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.691807 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.691838 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.691857 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.794675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.794829 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.794858 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.794939 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.794970 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.898247 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.898310 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.898327 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.898350 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:39 crc kubenswrapper[4956]: I1126 16:53:39.898366 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:39Z","lastTransitionTime":"2025-11-26T16:53:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.001051 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.001114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.001131 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.001156 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.001174 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.104416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.104517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.104545 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.104585 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.104615 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.207035 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.207095 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.207112 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.207136 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.207156 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.309453 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.309539 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.309567 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.309602 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.309629 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.413614 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.413670 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.413686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.413709 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.413722 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.517454 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.517510 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.517520 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.517541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.517553 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.620618 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.620674 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.620692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.620714 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.620728 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.724073 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.724153 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.724174 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.724214 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.724239 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.827178 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.827223 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.827236 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.827257 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.827272 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.930192 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.930258 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.930274 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.930303 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.930315 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:40Z","lastTransitionTime":"2025-11-26T16:53:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.995292 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.995340 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:40 crc kubenswrapper[4956]: E1126 16:53:40.995560 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.995657 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:40 crc kubenswrapper[4956]: E1126 16:53:40.995973 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:40 crc kubenswrapper[4956]: I1126 16:53:40.996352 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:40 crc kubenswrapper[4956]: E1126 16:53:40.996509 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:40 crc kubenswrapper[4956]: E1126 16:53:40.997261 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.033464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.033509 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.033521 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.033537 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.033549 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.136984 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.137054 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.137075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.137105 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.137126 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.240553 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.240623 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.240640 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.240670 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.240689 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.343841 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.343918 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.343936 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.343961 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.343999 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.447819 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.447882 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.447893 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.447915 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.447928 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.550487 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.550583 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.550611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.550643 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.550665 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.653793 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.653844 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.653857 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.653903 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.653918 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.756879 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.756933 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.756948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.756971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.756983 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.860553 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.860643 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.860674 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.860710 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.860736 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.963363 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.963453 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.963473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.963506 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:41 crc kubenswrapper[4956]: I1126 16:53:41.963528 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:41Z","lastTransitionTime":"2025-11-26T16:53:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.011456 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.066806 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.066889 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.066911 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.066940 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.066961 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.128325 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.128385 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.128397 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.128418 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.128434 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.141376 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:42Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.145287 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.145327 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.145339 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.145360 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.145371 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.159481 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:42Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.163625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.163682 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.163697 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.163711 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.163721 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.176563 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:42Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.181100 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.181145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.181163 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.181186 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.181208 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.195075 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:42Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.199328 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.199385 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.199395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.199413 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.199425 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.210889 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:42Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.211097 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.212721 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.212777 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.212787 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.212800 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.212809 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.315513 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.315615 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.315652 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.315676 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.315705 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.420389 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.420428 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.420438 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.420454 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.420464 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.524226 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.524282 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.524303 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.524328 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.524345 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.626351 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.626417 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.626448 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.626489 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.626504 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.729427 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.729486 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.729509 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.729534 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.729556 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.833494 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.833551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.833563 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.833588 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.833602 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.936290 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.936377 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.936415 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.936447 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.936471 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:42Z","lastTransitionTime":"2025-11-26T16:53:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.994498 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.994635 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.994498 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.994729 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:42 crc kubenswrapper[4956]: I1126 16:53:42.994776 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.994909 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.995101 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:42 crc kubenswrapper[4956]: E1126 16:53:42.995203 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.038790 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.038858 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.038894 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.038912 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.038927 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.142076 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.142147 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.142164 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.142193 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.142216 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.246117 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.246203 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.246225 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.246254 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.246273 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.349615 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.349683 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.349702 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.349731 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.349749 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.454191 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.454275 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.454295 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.454327 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.454353 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.557634 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.557707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.557726 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.557751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.557768 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.660030 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.660098 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.660110 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.660149 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.660159 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.762751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.762827 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.762847 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.762908 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.762926 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.866785 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.866891 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.866911 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.866942 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.866960 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.970673 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.970748 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.970757 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.970774 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:43 crc kubenswrapper[4956]: I1126 16:53:43.970786 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:43Z","lastTransitionTime":"2025-11-26T16:53:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.074494 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.074562 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.074580 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.074608 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.074626 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.177638 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.177721 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.177747 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.177782 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.177804 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.281482 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.281559 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.281578 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.281610 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.281631 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.385483 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.385541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.385557 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.385583 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.385600 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.489569 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.489645 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.489687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.489719 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.489739 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.593626 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.593690 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.593713 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.593741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.593762 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.696228 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.696270 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.696280 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.696296 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.696306 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.800264 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.800317 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.800327 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.800347 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.800361 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.902943 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.903024 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.903043 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.903074 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.903094 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:44Z","lastTransitionTime":"2025-11-26T16:53:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.994949 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.995012 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.995040 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:44 crc kubenswrapper[4956]: E1126 16:53:44.995121 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:44 crc kubenswrapper[4956]: I1126 16:53:44.995204 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:44 crc kubenswrapper[4956]: E1126 16:53:44.995435 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:44 crc kubenswrapper[4956]: E1126 16:53:44.995818 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:44 crc kubenswrapper[4956]: E1126 16:53:44.997101 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.006029 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.006074 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.006092 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.006116 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.006137 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.010956 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.022167 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.034065 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.048240 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.063949 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.080836 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.096913 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.109001 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.109049 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.109086 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.109106 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.109118 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.115829 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.155879 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.172764 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.188938 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.204430 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.212007 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.212160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.212306 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.212454 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.212564 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.223202 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.237664 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.253666 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.265165 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.280195 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.297523 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:45Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.315055 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.315266 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.315325 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.315398 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.315462 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.418184 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.418508 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.418571 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.418649 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.418712 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.521742 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.521800 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.521811 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.521833 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.521845 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.624989 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.625413 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.625507 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.625592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.625660 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.729176 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.729252 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.729272 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.729303 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.729330 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.833564 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.833627 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.833648 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.833671 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.833688 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.936988 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.937470 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.937625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.937778 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:45 crc kubenswrapper[4956]: I1126 16:53:45.937954 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:45Z","lastTransitionTime":"2025-11-26T16:53:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.042657 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.042726 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.042739 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.042769 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.042783 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.145651 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.145744 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.145771 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.145808 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.145836 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.250179 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.250269 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.250297 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.250333 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.250356 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.352973 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.353043 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.353063 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.353092 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.353112 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.456557 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.456621 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.456639 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.456667 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.456686 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.559381 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.559447 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.559466 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.559492 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.559511 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.661815 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.662114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.662319 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.662476 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.662627 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.767639 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.767705 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.767720 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.767740 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.767756 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.871898 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.871961 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.871974 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.871997 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.872010 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.975471 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.975531 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.975546 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.975572 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.975587 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:46Z","lastTransitionTime":"2025-11-26T16:53:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.994911 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.994948 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.994948 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:46 crc kubenswrapper[4956]: I1126 16:53:46.994963 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:46 crc kubenswrapper[4956]: E1126 16:53:46.995049 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:46 crc kubenswrapper[4956]: E1126 16:53:46.995200 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:46 crc kubenswrapper[4956]: E1126 16:53:46.995248 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:46 crc kubenswrapper[4956]: E1126 16:53:46.995328 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.079198 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.079291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.079312 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.079340 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.079359 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.182758 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.182824 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.182842 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.182902 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.182922 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.286149 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.286240 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.286260 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.286293 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.286320 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.368425 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:47 crc kubenswrapper[4956]: E1126 16:53:47.368781 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:47 crc kubenswrapper[4956]: E1126 16:53:47.369041 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:54:19.368972115 +0000 UTC m=+105.064932697 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.389471 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.389546 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.389566 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.389592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.389611 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.492682 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.492732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.492744 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.492764 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.492777 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.596100 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.596147 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.596156 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.596175 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.596185 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.699044 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.699555 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.699765 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.700022 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.700261 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.802998 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.803053 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.803066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.803086 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.803100 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.905608 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.905664 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.905675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.905693 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:47 crc kubenswrapper[4956]: I1126 16:53:47.905710 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:47Z","lastTransitionTime":"2025-11-26T16:53:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.009256 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.009357 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.009385 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.009424 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.009447 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.112675 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.112752 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.112802 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.112840 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.112932 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.217117 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.217228 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.217253 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.217287 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.217307 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.320199 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.320260 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.320273 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.320293 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.320311 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.423363 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.423732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.423797 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.423884 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.423958 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.527750 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.527851 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.527914 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.527964 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.527992 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.631299 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.631677 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.631758 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.631854 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.631973 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.734464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.734534 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.734551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.734573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.734588 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.838460 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.838517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.838533 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.838552 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.838564 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.941971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.942045 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.942059 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.942077 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.942091 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:48Z","lastTransitionTime":"2025-11-26T16:53:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.995275 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.995326 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.995289 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:48 crc kubenswrapper[4956]: I1126 16:53:48.995274 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:48 crc kubenswrapper[4956]: E1126 16:53:48.995500 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:48 crc kubenswrapper[4956]: E1126 16:53:48.995699 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:48 crc kubenswrapper[4956]: E1126 16:53:48.996013 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:48 crc kubenswrapper[4956]: E1126 16:53:48.996045 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.049509 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.049577 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.049606 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.049658 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.049681 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.153834 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.153972 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.153985 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.154007 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.154026 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.256226 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.256272 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.256281 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.256298 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.256311 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.359552 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.359612 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.359627 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.359652 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.359669 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.463272 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.463345 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.463368 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.463405 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.463430 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.541933 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/0.log" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.542039 4956 generic.go:334] "Generic (PLEG): container finished" podID="80ac0050-7f55-4845-ace7-a89f2442f884" containerID="7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4" exitCode=1 Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.542097 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerDied","Data":"7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.542714 4956 scope.go:117] "RemoveContainer" containerID="7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.563443 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.567460 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.567505 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.567517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.567540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.567553 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.584390 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:49Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.604140 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.623725 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.642561 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.657663 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.672618 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.672664 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.672680 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.672705 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.672721 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.676115 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.694255 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.710855 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.723062 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.744006 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.758596 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.774945 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.776090 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.776131 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.776145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.776164 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.776180 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.792577 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.806130 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.827217 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.846702 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.860112 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:49Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.879108 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.879188 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.879202 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.879227 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.879243 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.982146 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.982196 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.982207 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.982226 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:49 crc kubenswrapper[4956]: I1126 16:53:49.982237 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:49Z","lastTransitionTime":"2025-11-26T16:53:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.084953 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.084993 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.085003 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.085021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.085033 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.188484 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.188530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.188541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.188557 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.188568 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.291573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.291625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.291641 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.291663 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.291677 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.395361 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.395434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.395452 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.395483 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.395502 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.498984 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.499066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.499084 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.499113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.499133 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.548407 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/0.log" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.548486 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerStarted","Data":"bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.574989 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://bace073419edafb1239fce2bfe382c9b14f2919ed19618629fae402d6ebca8b8\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:13Z\\\",\\\"message\\\":\\\"tf after 0 failed attempt(s)\\\\nI1126 16:53:13.343038 6424 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:13.343041 6424 default_network_controller.go:776] Recording success event on pod openshift-network-operator/network-operator-58b4c7f79c-55gtf\\\\nI1126 16:53:13.342960 6424 obj_retry.go:365] Adding new object: *v1.Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf\\\\nI1126 16:53:13.343056 6424 ovn.go:134] Ensuring zone local for Pod openshift-network-diagnostics/network-check-source-55646444c4-trplf in node crc\\\\nI1126 16:53:13.342950 6424 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-additional-cni-plugins-vz5zh\\\\nI1126 16:53:13.343073 6424 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI1126 16:53:13.343102 6424 base_network_controller_pods.go:477] [default/openshift-network-diagnostics/network-check-source-55646444c4-trplf] creating logical port openshift-network-diagnostics_network-check-source-55646444c4-trplf for pod on switch crc\\\\nF1126 16:53:13.343149 6424 ovnkube.go:137] failed to run ovnkube: [failed to start network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:12Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.589148 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.602898 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.602963 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.602987 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.603022 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.603049 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.606945 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.626259 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.643543 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.666572 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.693194 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.706686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.706762 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.706783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.706815 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.706837 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.713812 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.737573 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.757939 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.776045 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.791303 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.809422 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.809473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.809492 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.809520 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.809538 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.809731 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.824901 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.840685 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.860880 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.875327 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.888485 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:50Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.913504 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.913577 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.913591 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.913618 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.913631 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:50Z","lastTransitionTime":"2025-11-26T16:53:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.995338 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.995391 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.995443 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:50 crc kubenswrapper[4956]: E1126 16:53:50.995590 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:50 crc kubenswrapper[4956]: I1126 16:53:50.995611 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:50 crc kubenswrapper[4956]: E1126 16:53:50.995758 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:50 crc kubenswrapper[4956]: E1126 16:53:50.996057 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:50 crc kubenswrapper[4956]: E1126 16:53:50.996309 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.017014 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.017089 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.017113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.017140 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.017161 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.120470 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.120532 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.120545 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.120569 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.120584 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.223406 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.223502 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.223537 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.223579 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.223608 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.328538 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.328623 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.328651 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.328686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.328710 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.432466 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.432544 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.432573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.432609 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.432635 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.537145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.537206 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.537223 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.537252 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.537269 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.640913 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.640985 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.641002 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.641028 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.641048 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.744361 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.744440 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.744460 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.744493 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.744512 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.848461 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.848542 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.848563 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.848593 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.848614 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.951555 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.951598 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.951611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.951628 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:51 crc kubenswrapper[4956]: I1126 16:53:51.951643 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:51Z","lastTransitionTime":"2025-11-26T16:53:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.054525 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.054604 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.054629 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.054660 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.054683 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.157610 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.157710 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.157735 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.157769 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.157789 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.261162 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.261248 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.261274 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.261346 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.261371 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.365373 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.365456 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.365485 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.365519 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.365537 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.420814 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.420920 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.420944 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.420975 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.420996 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.443672 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:52Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.449890 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.449945 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.449957 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.449980 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.449994 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.471060 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:52Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.476978 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.477046 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.477070 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.477100 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.477118 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.500746 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:52Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.506537 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.506611 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.506637 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.506681 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.506710 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.531020 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:52Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.536486 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.536521 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.536535 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.536559 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.536579 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.552530 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:52Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.552801 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.555760 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.555808 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.555821 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.555837 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.555850 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.659697 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.659762 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.659781 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.659809 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.659833 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.763923 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.763992 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.764014 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.764041 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.764060 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.868127 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.868187 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.868196 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.868216 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.868227 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.971288 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.971362 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.971384 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.971411 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.971429 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:52Z","lastTransitionTime":"2025-11-26T16:53:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.995028 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.995112 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.995132 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.995103 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.995823 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.996032 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.996259 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:52 crc kubenswrapper[4956]: I1126 16:53:52.996280 4956 scope.go:117] "RemoveContainer" containerID="72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b" Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.996444 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:52 crc kubenswrapper[4956]: E1126 16:53:52.996682 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.017781 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.043113 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.064891 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.074988 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.075075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.075102 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.075139 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.075170 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.084480 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.104184 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.125931 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.149798 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.173124 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.178687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.178764 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.178791 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.178828 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.178855 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.198292 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.229630 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.254649 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.274484 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.283665 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.283738 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.283755 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.283786 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.283806 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.298343 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.320365 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.337605 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.360055 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.378239 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.387203 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.387273 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.387291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.387316 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.387332 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.399238 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:53Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.490312 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.490360 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.490369 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.490384 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.490396 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.593698 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.593767 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.593782 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.593806 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.593824 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.697174 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.697243 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.697262 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.697287 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.697306 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.800337 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.800417 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.800435 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.800471 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.800495 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.904049 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.904133 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.904163 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.904199 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:53 crc kubenswrapper[4956]: I1126 16:53:53.904227 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:53Z","lastTransitionTime":"2025-11-26T16:53:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.008056 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.008130 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.008148 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.008181 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.008202 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.111747 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.111830 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.111853 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.111919 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.111945 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.215552 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.215630 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.215647 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.215680 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.215704 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.320272 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.320350 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.320371 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.320399 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.320418 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.423752 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.423827 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.423846 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.423912 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.423938 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.527444 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.527553 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.527584 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.527633 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.527665 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.631122 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.631179 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.631198 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.631224 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.631243 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.734618 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.734685 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.734704 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.734728 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.734743 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.838173 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.838264 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.838284 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.838311 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.838330 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.942840 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.942963 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.942985 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.943015 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.943036 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:54Z","lastTransitionTime":"2025-11-26T16:53:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.995280 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.995296 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.995421 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:54 crc kubenswrapper[4956]: E1126 16:53:54.995421 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:54 crc kubenswrapper[4956]: E1126 16:53:54.995510 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:54 crc kubenswrapper[4956]: I1126 16:53:54.995623 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:54 crc kubenswrapper[4956]: E1126 16:53:54.995717 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:54 crc kubenswrapper[4956]: E1126 16:53:54.995819 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.019650 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.044784 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.047665 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.047707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.047719 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.047737 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.047750 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.069077 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.090293 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.109607 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.130425 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.150189 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.151065 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.151126 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.151146 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.151176 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.151197 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.167368 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.184627 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.198908 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.228284 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.243026 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.254538 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.254595 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.254612 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.254639 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.254659 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.263388 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.282216 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.297467 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.320241 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.341496 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.356889 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:53:55Z is after 2025-08-24T17:21:41Z" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.357749 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.357821 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.357844 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.357905 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.357930 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.460855 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.461021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.461050 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.461084 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.461112 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.564929 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.565353 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.565408 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.565444 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.565471 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.668781 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.668820 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.668830 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.668846 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.668858 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.772451 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.772508 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.772526 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.772552 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.772567 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.875241 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.875301 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.875316 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.875337 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.875353 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.978208 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.978291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.978313 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.978346 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:55 crc kubenswrapper[4956]: I1126 16:53:55.978373 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:55Z","lastTransitionTime":"2025-11-26T16:53:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.082687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.082771 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.082851 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.082909 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.082925 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.187060 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.187114 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.187127 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.187144 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.187158 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.290464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.290559 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.290592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.290630 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.290683 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.395221 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.395299 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.395330 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.395365 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.395393 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.498662 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.498741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.498766 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.498801 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.498825 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.601825 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.601909 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.601927 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.601955 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.601974 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.705291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.705354 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.705375 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.705399 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.705415 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.808343 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.808397 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.808406 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.808464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.808478 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.911430 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.911506 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.911530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.911564 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.911588 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:56Z","lastTransitionTime":"2025-11-26T16:53:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.994554 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.994657 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.994554 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:56 crc kubenswrapper[4956]: I1126 16:53:56.994958 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:56 crc kubenswrapper[4956]: E1126 16:53:56.995129 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:56 crc kubenswrapper[4956]: E1126 16:53:56.995294 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:56 crc kubenswrapper[4956]: E1126 16:53:56.995444 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:56 crc kubenswrapper[4956]: E1126 16:53:56.995507 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.014005 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.014075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.014093 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.014117 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.014136 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.117546 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.117621 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.117645 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.117677 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.117702 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.220623 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.220684 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.220701 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.220726 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.220745 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.323565 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.323635 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.323654 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.323682 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.323701 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.427445 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.427530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.427549 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.427582 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.427606 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.532158 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.532269 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.532289 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.532324 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.532344 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.635668 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.635733 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.635746 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.635768 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.635781 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.739720 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.739787 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.739814 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.739850 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.739907 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.843751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.843835 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.843901 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.843940 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.843962 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.954122 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.954680 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.954701 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.954729 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:57 crc kubenswrapper[4956]: I1126 16:53:57.954748 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:57Z","lastTransitionTime":"2025-11-26T16:53:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.058855 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.058955 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.058974 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.059004 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.059028 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.163374 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.163848 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.164058 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.164264 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.164479 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.267686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.267761 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.267779 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.267807 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.267825 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.371318 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.371410 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.371435 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.371467 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.371495 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.475093 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.475160 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.475170 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.475188 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.475201 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.579464 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.579524 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.579540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.579563 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.579579 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.682940 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.682990 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.683000 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.683016 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.683027 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.786389 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.786449 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.786463 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.786483 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.786495 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.821643 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.821774 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.821827 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:02.821804068 +0000 UTC m=+148.517764620 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.821855 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.821922 4956 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.821957 4956 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.821968 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:55:02.821957443 +0000 UTC m=+148.517917985 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.821922 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.821985 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-11-26 16:55:02.821978153 +0000 UTC m=+148.517938705 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.822159 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822404 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822451 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822476 4956 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822588 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-11-26 16:55:02.822559339 +0000 UTC m=+148.518519931 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822924 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822960 4956 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.822980 4956 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.823032 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-11-26 16:55:02.823023442 +0000 UTC m=+148.518983994 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.891182 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.891296 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.891324 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.891369 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.891398 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995091 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995248 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.995273 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995503 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995543 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.995498 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995565 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995628 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995650 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:58Z","lastTransitionTime":"2025-11-26T16:53:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995694 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:53:58 crc kubenswrapper[4956]: I1126 16:53:58.995758 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.996034 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:53:58 crc kubenswrapper[4956]: E1126 16:53:58.996207 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.100028 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.100102 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.100121 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.100150 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.100170 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.203488 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.203531 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.203539 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.203554 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.203564 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.305859 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.305949 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.305968 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.305993 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.306010 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.409405 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.409483 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.409501 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.409531 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.409550 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.512064 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.512126 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.512144 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.512181 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.512222 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.615680 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.615763 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.615789 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.615824 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.615856 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.720075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.720163 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.720182 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.720213 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.720234 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.824572 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.824651 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.824668 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.824698 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.824729 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.928585 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.928644 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.928661 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.928680 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:53:59 crc kubenswrapper[4956]: I1126 16:53:59.928695 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:53:59Z","lastTransitionTime":"2025-11-26T16:53:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.032294 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.032388 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.032416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.032455 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.032482 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.136139 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.136203 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.136217 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.136238 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.136251 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.238847 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.238904 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.238914 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.238927 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.238938 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.342080 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.342134 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.342145 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.342163 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.342179 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.446473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.446868 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.447109 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.447245 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.447376 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.551326 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.551408 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.551429 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.551468 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.551489 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.654597 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.654680 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.654707 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.654740 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.654760 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.758075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.758151 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.758171 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.758200 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.758255 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.861722 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.861813 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.861832 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.861867 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.862043 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.964686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.964763 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.964775 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.964798 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.964819 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:00Z","lastTransitionTime":"2025-11-26T16:54:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.995243 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.995353 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.995376 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:00 crc kubenswrapper[4956]: E1126 16:54:00.995556 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:00 crc kubenswrapper[4956]: I1126 16:54:00.995578 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:00 crc kubenswrapper[4956]: E1126 16:54:00.995715 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:00 crc kubenswrapper[4956]: E1126 16:54:00.995852 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:00 crc kubenswrapper[4956]: E1126 16:54:00.996005 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.068366 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.068426 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.068440 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.068466 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.068483 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.172223 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.172308 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.172322 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.172340 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.172357 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.275423 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.275486 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.275506 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.275530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.275549 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.378687 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.378754 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.378768 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.378791 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.378807 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.481449 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.481493 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.481505 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.481521 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.481531 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.584472 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.584535 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.584552 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.584573 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.584589 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.687434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.687500 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.687517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.687540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.687556 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.790847 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.790921 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.790932 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.790953 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.790965 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.893842 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.893904 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.893920 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.893936 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.893947 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.997451 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.997522 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.997539 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.997562 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:01 crc kubenswrapper[4956]: I1126 16:54:01.997576 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:01Z","lastTransitionTime":"2025-11-26T16:54:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.100414 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.100461 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.100472 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.100493 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.100504 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.204062 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.204117 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.204128 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.204148 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.204162 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.307811 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.307873 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.307920 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.307945 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.307967 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.411416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.411482 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.411500 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.411527 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.411548 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.514858 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.514959 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.514982 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.515010 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.515030 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.618066 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.618148 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.618166 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.618195 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.618215 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.721133 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.721198 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.721218 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.721245 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.721268 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.825274 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.825346 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.825362 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.825387 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.825405 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.906314 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.906403 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.906425 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.906451 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.906468 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.928671 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.934332 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.934368 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.934379 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.934400 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.934412 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.956696 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.962312 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.962350 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.962360 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.962374 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.962383 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.987024 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:02Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.992858 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.992950 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.992971 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.992998 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.993019 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:02Z","lastTransitionTime":"2025-11-26T16:54:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.994483 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.994704 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.994771 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:02 crc kubenswrapper[4956]: I1126 16:54:02.994764 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.994949 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.995095 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.995245 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:02 crc kubenswrapper[4956]: E1126 16:54:02.995359 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:03 crc kubenswrapper[4956]: E1126 16:54:03.015656 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.020709 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.020763 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.020786 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.020814 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.020836 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: E1126 16:54:03.045113 4956 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"lastTransitionTime\\\":\\\"2025-11-26T16:54:03Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"fcb74053-066d-43b8-98e1-af830e827a8c\\\",\\\"systemUUID\\\":\\\"4a4a5395-957e-4512-a554-c3c322e283ff\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:03Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:03 crc kubenswrapper[4956]: E1126 16:54:03.045429 4956 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.048348 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.048403 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.048422 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.048446 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.048464 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.151871 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.151960 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.151976 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.152000 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.152017 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.254931 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.255004 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.255021 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.255047 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.255068 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.358227 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.358306 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.358328 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.358360 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.358384 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.461378 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.461459 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.461486 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.461517 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.461541 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.565402 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.565503 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.565520 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.565548 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.565574 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.668592 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.668696 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.668715 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.668742 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.668763 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.772019 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.772123 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.772148 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.772183 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.772209 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.875488 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.875555 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.875572 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.875603 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.875624 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.979439 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.979547 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.979669 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.979708 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.979736 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:03Z","lastTransitionTime":"2025-11-26T16:54:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:03 crc kubenswrapper[4956]: I1126 16:54:03.995807 4956 scope.go:117] "RemoveContainer" containerID="72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.083935 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.084063 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.084129 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.084202 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.084235 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.187468 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.187935 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.187946 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.187961 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.187972 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.290866 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.290925 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.290936 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.290952 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.290965 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.393560 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.393595 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.393604 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.393622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.393633 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.518015 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.518068 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.518079 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.518096 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.518105 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.607332 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/2.log" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.610845 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.611283 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.620188 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.620222 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.620233 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.620248 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.620258 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.627111 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.637910 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.647578 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.664851 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.686157 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.696582 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.711105 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.724250 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.724317 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.724341 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.724371 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.724486 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.725601 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.739445 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.754346 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.771844 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.791824 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.811985 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.827200 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.827253 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.827422 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.827433 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.827450 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.827459 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.842104 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.855896 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.868021 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.879706 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:04Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.930768 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.930852 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.930888 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.930910 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.930928 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:04Z","lastTransitionTime":"2025-11-26T16:54:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.994681 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:04 crc kubenswrapper[4956]: E1126 16:54:04.994809 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.994705 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:04 crc kubenswrapper[4956]: E1126 16:54:04.994913 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.994688 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:04 crc kubenswrapper[4956]: E1126 16:54:04.994963 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:04 crc kubenswrapper[4956]: I1126 16:54:04.995250 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:04 crc kubenswrapper[4956]: E1126 16:54:04.995295 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.012330 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.026074 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.033770 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.033848 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.033893 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.033918 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.033942 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.041422 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.055739 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.072129 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.087137 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.102621 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.115743 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.130973 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.135578 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.135626 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.135636 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.135657 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.135670 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.143668 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.158718 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.173147 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.196846 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.221250 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.233831 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.237809 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.237853 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.237883 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.237898 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.237908 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.250818 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.268850 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.282084 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.340767 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.340817 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.340829 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.340850 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.340887 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.443645 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.443714 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.443744 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.443777 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.443801 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.547154 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.547214 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.547235 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.547266 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.547287 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.618522 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/3.log" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.619457 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/2.log" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.624147 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" exitCode=1 Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.624196 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.624234 4956 scope.go:117] "RemoveContainer" containerID="72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.626087 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 16:54:05 crc kubenswrapper[4956]: E1126 16:54:05.627229 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.651478 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.651523 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.651535 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.651551 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.651562 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.663999 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://72a2412c6bfecc678b8baea39968ca3ac4097dcb0818e7246b7ce9ef711f125b\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:37Z\\\",\\\"message\\\":\\\"19615025667110816) with []\\\\nI1126 16:53:37.432730 6708 address_set.go:302] New(bf133528-8652-4c84-85ff-881f0afe9837/default-network-controller:EgressService:egresssvc-served-pods:v4/a13607449821398607916) with []\\\\nI1126 16:53:37.432894 6708 factory.go:1336] Added *v1.Node event handler 7\\\\nI1126 16:53:37.432937 6708 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI1126 16:53:37.433046 6708 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI1126 16:53:37.433072 6708 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI1126 16:53:37.433149 6708 handler.go:208] Removed *v1.Node event handler 7\\\\nI1126 16:53:37.433179 6708 handler.go:208] Removed *v1.Node event handler 2\\\\nI1126 16:53:37.433374 6708 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.433938 6708 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI1126 16:53:37.434000 6708 factory.go:656] Stopping watch factory\\\\nI1126 16:53:37.434107 6708 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI1126 16:53:37.434658 6708 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI1126 16:53:37.434825 6708 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:53:37.434853 6708 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:53:37.434949 6708 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:36Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:54:04Z\\\",\\\"message\\\":\\\"sFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 16:54:04.948277 7057 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1126 16:54:04.948828 7057 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:54:04.948940 7057 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 16:54:04.949004 7057 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 16:54:04.965391 7057 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1126 16:54:04.965419 7057 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1126 16:54:04.965507 7057 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:54:04.965554 7057 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:54:04.965653 7057 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:54:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.680074 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.698454 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.719273 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.734545 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.752196 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.755505 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.755631 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.755703 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.755741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.755984 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.769749 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.783016 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.799272 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.814113 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.833839 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.852507 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.859401 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.859466 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.859488 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.859518 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.859540 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.916659 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.934576 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.950260 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.962462 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.962502 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.962513 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.962534 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.962547 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:05Z","lastTransitionTime":"2025-11-26T16:54:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.968047 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:05 crc kubenswrapper[4956]: I1126 16:54:05.984572 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.001670 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:05Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.067062 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.067141 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.067165 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.067194 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.067214 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.170684 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.170740 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.170759 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.170786 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.170806 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.273597 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.273692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.273711 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.273737 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.273754 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.377705 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.377843 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.377915 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.377946 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.378003 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.481715 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.481792 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.481816 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.481848 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.481899 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.584905 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.584989 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.585011 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.585042 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.585064 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.632008 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/3.log" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.638369 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 16:54:06 crc kubenswrapper[4956]: E1126 16:54:06.638747 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.658916 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://578753f2e809cd2085d57cc367844992021d063220d861bf289e5d61eb92cf6a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.680360 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:56Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1466a7c6556e297d3a5d1b9d027af65b0f4ada8501e80d568afe25011b6066b5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://31e7cdcc37cda83cf30832d088f27299f5e3f2786cc8b36875c09fa681abf2d6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.688239 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.688310 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.688332 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.688357 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.688376 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.699391 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-vhvtw" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"adc2efe4-3ec6-4e5f-8dd8-76954150acb1\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8898470790e73ad6c62c0a92839f95d4939703b16913b13aafb779eea7e0d228\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8dtp\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-vhvtw\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.726213 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d94ea852-698d-41f1-93e4-df0322a23b47\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:09Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://606c45bad954807da485adb932e9bf8400040ec38f51c2265f3b8c4c2e1917c6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://47de7e2dd41a04360e1b260eb56d459fae1f53b3bdbad53e3eaaa197507b7a3b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4f3cc341abbe26588940ae8b3f1d7d3647149b469e83a4e6798d014505f83709\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b966b95d95ba59bcfc4203b47b5b45bd2a9d8257986688d9bc5b0f796a27e9d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:04Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:04Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://44373f9dcd46f26e2a3de3ad0ea68c85b6ef9f502e4f86baf5dd0ec558a706da\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:05Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:05Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://948849c34f6da2441091430d8421d912cdc67bf5d2363bd865e3ed817cea7668\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:06Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f83976975d6eb8acc52f703ab577f51a66675f314c20574d99fa031eb43900ba\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:07Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-bxgn8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-vz5zh\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.752088 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"46bd4702-b565-4452-be2c-36a3f1f48621\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:03Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:54:04Z\\\",\\\"message\\\":\\\"sFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 16:54:04.948277 7057 reflector.go:311] Stopping reflector *v1.NetworkAttachmentDefinition (0s) from github.com/k8snetworkplumbingwg/network-attachment-definition-client/pkg/client/informers/externalversions/factory.go:117\\\\nI1126 16:54:04.948828 7057 reflector.go:311] Stopping reflector *v1.Pod (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1126 16:54:04.948940 7057 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 16:54:04.949004 7057 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1126 16:54:04.965391 7057 shared_informer.go:320] Caches are synced for node-tracker-controller\\\\nI1126 16:54:04.965419 7057 services_controller.go:204] Setting up event handlers for services for network=default\\\\nI1126 16:54:04.965507 7057 ovnkube.go:599] Stopped ovnkube\\\\nI1126 16:54:04.965554 7057 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nF1126 16:54:04.965653 7057 ovnkube.go:137] failed to run ov\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:54:04Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-zzdlc\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:02Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f2thc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.770141 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-vpszs" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8dcb7d83-aba3-4ec1-83f2-2afd638d3911\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:07Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:08Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://138ce971ec3b85f434f38ac53ea5b6d078dd59c00702f1d6ecb49fed47162365\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:08Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-924q5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:07Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-vpszs\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.791953 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.792034 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.792054 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.792083 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.792105 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.799249 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1c5fd9ff-5bc2-49ef-a042-8042c97467ee\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:10Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"ing back to namespace): Get \\\\\\\"https://localhost:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\\\\\\\": net/http: TLS handshake timeout\\\\nI1126 16:52:48.686411 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI1126 16:52:48.688909 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-3631915393/tls.crt::/tmp/serving-cert-3631915393/tls.key\\\\\\\"\\\\nI1126 16:52:54.772850 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI1126 16:52:54.779322 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI1126 16:52:54.779374 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI1126 16:52:54.779427 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI1126 16:52:54.779444 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1126 16:52:54.791175 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1126 16:52:54.791228 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791239 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1126 16:52:54.791251 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW1126 16:52:54.791259 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1126 16:52:54.791267 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1126 16:52:54.791275 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1126 16:52:54.791407 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nF1126 16:52:54.797381 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:55Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:38Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.812159 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"13daad8e-bbbd-4481-bc0c-8047faf3900e\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://772dedbe5880cab0799834216789dae0b64ecfbb1a2210475e07b25913de6057\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://30eceb6934e3180610f79d9e5fd48b81e6263237996d160bcf0d2b3fd12587d5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.827328 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"20f2376e-efa2-4242-a756-3059e6f9d460\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49be1350608e3955060f313e2108b0ad92a4299de720ff5471e680474a52c19a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d59282628163874600e8be48f982200b55811d8a82f856a7997ce3c558877019\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1e233a865caf59e114d9e763a2582bc59573518c7b57bd64f281165a2f2aeb69\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.845645 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.864358 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.880728 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:59Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ba0af52c6d8af2ad06b4a0e9fccde5eb2c31be348a0c36e78a6bd5e429a08d96\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:58Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.895434 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.895473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.895504 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.895933 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.895959 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.897036 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:54Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.914653 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-m4dm4" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"80ac0050-7f55-4845-ace7-a89f2442f884\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:03Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:50Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-11-26T16:53:48Z\\\",\\\"message\\\":\\\"2025-11-26T16:53:03+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3\\\\n2025-11-26T16:53:03+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_8a07da7b-24dd-4db2-b16b-892c458a87a3 to /host/opt/cni/bin/\\\\n2025-11-26T16:53:03Z [verbose] multus-daemon started\\\\n2025-11-26T16:53:03Z [verbose] Readiness Indicator file check\\\\n2025-11-26T16:53:48Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-11-26T16:53:02Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:49Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-85hhq\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-multus\"/\"multus-m4dm4\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.930110 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c8f833d9-cb7c-4afc-ac70-019f6a1e2d30\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:37Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:29Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:52:35Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://833c61aa4f7671158003cc4c0e04fd15e4462c31ed5cf3b1627abb7c5d8cfca5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://408fd95d0ce43fdba96cff557947fd3fb594e05e191cda98f1e44f7b1bea7c8e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://acff41124547070203c92ec0983421eafec9cdf33b7fbce729da6282cc8ab855\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:52:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0c72fe125b028d4864d815a04da098ebb628b0193e7d00c388ba10c663477209\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-11-26T16:52:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-11-26T16:52:36Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:52:35Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.945821 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"fadaf3cf-cfc5-4f60-bd4a-4eae814da018\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:01Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://364c49d069f57dc4d480cb3478dfad1d9c9d37193070624515bfed3e68e36b45\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-jrmtl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:01Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-9rvrb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.959264 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8bfae4c3-3d62-4968-91d4-8a8caabc4af8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://27d938365356cfefc302d7d2de1d38addfe12f5d60d7d98f4d7f4fcf295a48d1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://884e3c1fb1c7b0adbec1ca9fb60d7073521c51cca3c65f24b1b41f734bbcd6e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-11-26T16:53:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7zg7\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-f6knr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.973561 4956 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"0956e12e-96b8-4b83-bc0a-666639b36586\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-11-26T16:53:15Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6nc92\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-11-26T16:53:15Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-b8nt2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-11-26T16:54:06Z is after 2025-08-24T17:21:41Z" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.995136 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:06 crc kubenswrapper[4956]: E1126 16:54:06.995399 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.995438 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.995501 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.995496 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:06 crc kubenswrapper[4956]: E1126 16:54:06.995715 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:06 crc kubenswrapper[4956]: E1126 16:54:06.995768 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:06 crc kubenswrapper[4956]: E1126 16:54:06.995837 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.998040 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.998078 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.998089 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.998103 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:06 crc kubenswrapper[4956]: I1126 16:54:06.998116 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:06Z","lastTransitionTime":"2025-11-26T16:54:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.101342 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.101395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.101406 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.101425 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.101436 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.204169 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.204239 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.204256 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.204281 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.204299 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.307732 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.307838 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.307858 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.307916 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.307937 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.411735 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.411794 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.411810 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.411834 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.411917 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.515526 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.515605 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.515637 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.515666 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.515688 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.619193 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.619243 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.619254 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.619274 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.619288 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.722187 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.722255 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.722272 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.722294 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.722309 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.825668 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.825741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.825759 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.825788 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.825809 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.930071 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.930151 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.930175 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.930207 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:07 crc kubenswrapper[4956]: I1126 16:54:07.930233 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:07Z","lastTransitionTime":"2025-11-26T16:54:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.032541 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.032597 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.032606 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.032650 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.032662 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.136051 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.136125 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.136167 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.136209 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.136233 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.240327 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.240399 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.240427 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.240478 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.240505 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.344412 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.344495 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.344518 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.344548 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.344569 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.447841 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.447988 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.448005 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.448031 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.448048 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.551097 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.551143 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.551154 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.551172 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.551183 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.654789 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.654851 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.654898 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.654924 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.654944 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.757538 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.757595 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.757610 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.757629 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.757640 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.860655 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.860748 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.860771 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.860801 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.860823 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.965067 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.965124 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.965135 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.965154 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.965170 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:08Z","lastTransitionTime":"2025-11-26T16:54:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.995242 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.995349 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:08 crc kubenswrapper[4956]: E1126 16:54:08.995447 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.995897 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:08 crc kubenswrapper[4956]: E1126 16:54:08.996206 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:08 crc kubenswrapper[4956]: I1126 16:54:08.996316 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:08 crc kubenswrapper[4956]: E1126 16:54:08.996568 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:08 crc kubenswrapper[4956]: E1126 16:54:08.996326 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.069356 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.069737 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.069824 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.069948 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.070086 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.172957 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.173009 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.173020 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.173041 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.173055 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.276753 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.276807 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.276818 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.276837 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.276854 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.380447 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.380511 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.380522 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.380540 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.380552 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.483621 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.483692 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.483709 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.483734 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.483753 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.587413 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.587501 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.587525 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.587554 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.587578 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.690802 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.690946 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.690968 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.690993 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.691010 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.794409 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.794473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.794494 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.794526 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.794546 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.897346 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.897413 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.897428 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.897452 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:09 crc kubenswrapper[4956]: I1126 16:54:09.897468 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:09Z","lastTransitionTime":"2025-11-26T16:54:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.000201 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.000274 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.000290 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.000313 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.000330 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.104024 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.104070 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.104087 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.104113 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.104132 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.207830 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.207911 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.207928 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.207949 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.207964 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.311186 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.311227 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.311242 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.311257 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.311270 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.414665 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.414718 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.414741 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.414760 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.414774 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.516981 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.517048 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.517075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.517107 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.517129 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.620734 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.620774 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.620788 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.620807 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.620820 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.724309 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.724392 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.724412 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.724444 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.724467 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.827909 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.828008 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.828039 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.828077 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.828109 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.931840 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.932430 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.932451 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.932481 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.932503 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:10Z","lastTransitionTime":"2025-11-26T16:54:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.994979 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.995082 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:10 crc kubenswrapper[4956]: E1126 16:54:10.995183 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.995332 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:10 crc kubenswrapper[4956]: I1126 16:54:10.995386 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:10 crc kubenswrapper[4956]: E1126 16:54:10.995879 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:10 crc kubenswrapper[4956]: E1126 16:54:10.996021 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:10 crc kubenswrapper[4956]: E1126 16:54:10.996272 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.036997 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.037078 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.037092 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.037112 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.037124 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.140672 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.140751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.140776 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.140806 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.140827 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.244101 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.244174 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.244198 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.244226 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.244244 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.347007 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.347065 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.347082 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.347105 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.347119 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.450730 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.450807 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.450827 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.450857 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.450906 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.553436 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.553501 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.553530 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.553562 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.553584 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.656566 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.656622 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.656638 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.656663 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.656680 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.760727 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.760792 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.760813 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.760841 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.760861 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.863633 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.863686 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.863698 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.863716 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.863729 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.967125 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.967195 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.967215 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.967244 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:11 crc kubenswrapper[4956]: I1126 16:54:11.967266 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:11Z","lastTransitionTime":"2025-11-26T16:54:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.070320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.070392 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.070411 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.070446 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.070473 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.174291 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.174358 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.174395 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.174468 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.174492 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.277786 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.277859 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.277907 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.277934 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.277955 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.381662 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.381757 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.381783 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.381842 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.381862 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.485328 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.485416 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.485433 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.485459 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.485477 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.589247 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.589294 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.589309 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.589333 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.589355 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.692963 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.693055 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.693075 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.693101 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.693152 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.796625 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.796695 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.796715 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.796744 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.796765 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.900320 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.900453 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.900473 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.900498 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.900515 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:12Z","lastTransitionTime":"2025-11-26T16:54:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.995187 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.995276 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.995340 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:12 crc kubenswrapper[4956]: I1126 16:54:12.995361 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:12 crc kubenswrapper[4956]: E1126 16:54:12.995364 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:12 crc kubenswrapper[4956]: E1126 16:54:12.995459 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:12 crc kubenswrapper[4956]: E1126 16:54:12.995606 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:12 crc kubenswrapper[4956]: E1126 16:54:12.995827 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.003758 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.003818 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.003837 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.003889 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.003910 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:13Z","lastTransitionTime":"2025-11-26T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.106972 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.107015 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.107024 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.107037 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.107048 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:13Z","lastTransitionTime":"2025-11-26T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.161751 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.161786 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.161795 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.161810 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.161818 4956 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-11-26T16:54:13Z","lastTransitionTime":"2025-11-26T16:54:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.243453 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578"] Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.244031 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.246931 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.247395 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.247492 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.248949 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.320262 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-vz5zh" podStartSLOduration=73.32023855 podStartE2EDuration="1m13.32023855s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.28069009 +0000 UTC m=+98.976650682" watchObservedRunningTime="2025-11-26 16:54:13.32023855 +0000 UTC m=+99.016199142" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.363780 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-vpszs" podStartSLOduration=73.36375175 podStartE2EDuration="1m13.36375175s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.332843091 +0000 UTC m=+99.028803713" watchObservedRunningTime="2025-11-26 16:54:13.36375175 +0000 UTC m=+99.059712342" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.398610 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-vhvtw" podStartSLOduration=73.398578819 podStartE2EDuration="1m13.398578819s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.397595352 +0000 UTC m=+99.093555914" watchObservedRunningTime="2025-11-26 16:54:13.398578819 +0000 UTC m=+99.094539421" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.420703 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=78.420686034 podStartE2EDuration="1m18.420686034s" podCreationTimestamp="2025-11-26 16:52:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.420250602 +0000 UTC m=+99.116211194" watchObservedRunningTime="2025-11-26 16:54:13.420686034 +0000 UTC m=+99.116646586" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.422940 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/526bbd8d-715a-4335-a309-6cde08b9483f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.423071 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/526bbd8d-715a-4335-a309-6cde08b9483f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.423227 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/526bbd8d-715a-4335-a309-6cde08b9483f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.423320 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526bbd8d-715a-4335-a309-6cde08b9483f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.423379 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/526bbd8d-715a-4335-a309-6cde08b9483f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.439215 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=31.439186539 podStartE2EDuration="31.439186539s" podCreationTimestamp="2025-11-26 16:53:42 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.439066945 +0000 UTC m=+99.135027497" watchObservedRunningTime="2025-11-26 16:54:13.439186539 +0000 UTC m=+99.135147131" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.496961 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-m4dm4" podStartSLOduration=73.496937195 podStartE2EDuration="1m13.496937195s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.496257666 +0000 UTC m=+99.192218258" watchObservedRunningTime="2025-11-26 16:54:13.496937195 +0000 UTC m=+99.192897797" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.522194 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=79.522170267 podStartE2EDuration="1m19.522170267s" podCreationTimestamp="2025-11-26 16:52:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.521821087 +0000 UTC m=+99.217781709" watchObservedRunningTime="2025-11-26 16:54:13.522170267 +0000 UTC m=+99.218130849" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.524553 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/526bbd8d-715a-4335-a309-6cde08b9483f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.524644 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/526bbd8d-715a-4335-a309-6cde08b9483f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.524737 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/526bbd8d-715a-4335-a309-6cde08b9483f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.524829 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/526bbd8d-715a-4335-a309-6cde08b9483f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.524934 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/526bbd8d-715a-4335-a309-6cde08b9483f-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.524955 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526bbd8d-715a-4335-a309-6cde08b9483f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.525059 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/526bbd8d-715a-4335-a309-6cde08b9483f-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.526066 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/526bbd8d-715a-4335-a309-6cde08b9483f-service-ca\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.533752 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/526bbd8d-715a-4335-a309-6cde08b9483f-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.557956 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/526bbd8d-715a-4335-a309-6cde08b9483f-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-nr578\" (UID: \"526bbd8d-715a-4335-a309-6cde08b9483f\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.570087 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.634127 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=44.63409422 podStartE2EDuration="44.63409422s" podCreationTimestamp="2025-11-26 16:53:29 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.614591657 +0000 UTC m=+99.310552209" watchObservedRunningTime="2025-11-26 16:54:13.63409422 +0000 UTC m=+99.330054812" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.634835 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podStartSLOduration=73.6348193 podStartE2EDuration="1m13.6348193s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.633628947 +0000 UTC m=+99.329589529" watchObservedRunningTime="2025-11-26 16:54:13.6348193 +0000 UTC m=+99.330779892" Nov 26 16:54:13 crc kubenswrapper[4956]: I1126 16:54:13.663943 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" event={"ID":"526bbd8d-715a-4335-a309-6cde08b9483f","Type":"ContainerStarted","Data":"12fcd4e7577af1e4bb772d73b0dfd8c69f7617c876447ab5c6c5622fa2bd5e79"} Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.669735 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" event={"ID":"526bbd8d-715a-4335-a309-6cde08b9483f","Type":"ContainerStarted","Data":"23ec3c32451b1c1496221a031b8b618dd94a2c9e6bd3f099dbe8fba907a177d1"} Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.692902 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-nr578" podStartSLOduration=74.692883109 podStartE2EDuration="1m14.692883109s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:14.692814897 +0000 UTC m=+100.388775519" watchObservedRunningTime="2025-11-26 16:54:14.692883109 +0000 UTC m=+100.388843661" Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.693031 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-f6knr" podStartSLOduration=74.693026453 podStartE2EDuration="1m14.693026453s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:13.650829195 +0000 UTC m=+99.346789757" watchObservedRunningTime="2025-11-26 16:54:14.693026453 +0000 UTC m=+100.388987005" Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.994333 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:14 crc kubenswrapper[4956]: E1126 16:54:14.996101 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.996204 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.996301 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:14 crc kubenswrapper[4956]: I1126 16:54:14.996335 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:14 crc kubenswrapper[4956]: E1126 16:54:14.996425 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:14 crc kubenswrapper[4956]: E1126 16:54:14.996565 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:14 crc kubenswrapper[4956]: E1126 16:54:14.996670 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:16 crc kubenswrapper[4956]: I1126 16:54:16.994992 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:16 crc kubenswrapper[4956]: I1126 16:54:16.995072 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:16 crc kubenswrapper[4956]: I1126 16:54:16.995179 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:16 crc kubenswrapper[4956]: I1126 16:54:16.995303 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:16 crc kubenswrapper[4956]: E1126 16:54:16.996947 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:16 crc kubenswrapper[4956]: E1126 16:54:16.997191 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:16 crc kubenswrapper[4956]: E1126 16:54:16.997334 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:16 crc kubenswrapper[4956]: E1126 16:54:16.997496 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:16 crc kubenswrapper[4956]: I1126 16:54:16.997924 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 16:54:16 crc kubenswrapper[4956]: E1126 16:54:16.998301 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:54:18 crc kubenswrapper[4956]: I1126 16:54:18.994986 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:18 crc kubenswrapper[4956]: I1126 16:54:18.994998 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:18 crc kubenswrapper[4956]: I1126 16:54:18.995036 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:18 crc kubenswrapper[4956]: E1126 16:54:18.995994 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:18 crc kubenswrapper[4956]: I1126 16:54:18.995327 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:18 crc kubenswrapper[4956]: E1126 16:54:18.996122 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:18 crc kubenswrapper[4956]: E1126 16:54:18.996195 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:18 crc kubenswrapper[4956]: E1126 16:54:18.995756 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:19 crc kubenswrapper[4956]: I1126 16:54:19.400191 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:19 crc kubenswrapper[4956]: E1126 16:54:19.400451 4956 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:54:19 crc kubenswrapper[4956]: E1126 16:54:19.400529 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs podName:0956e12e-96b8-4b83-bc0a-666639b36586 nodeName:}" failed. No retries permitted until 2025-11-26 16:55:23.400504222 +0000 UTC m=+169.096464814 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs") pod "network-metrics-daemon-b8nt2" (UID: "0956e12e-96b8-4b83-bc0a-666639b36586") : object "openshift-multus"/"metrics-daemon-secret" not registered Nov 26 16:54:20 crc kubenswrapper[4956]: I1126 16:54:20.023202 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Nov 26 16:54:20 crc kubenswrapper[4956]: I1126 16:54:20.995318 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:20 crc kubenswrapper[4956]: I1126 16:54:20.995356 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:20 crc kubenswrapper[4956]: I1126 16:54:20.995446 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:20 crc kubenswrapper[4956]: E1126 16:54:20.995566 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:20 crc kubenswrapper[4956]: E1126 16:54:20.995722 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:20 crc kubenswrapper[4956]: E1126 16:54:20.995971 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:20 crc kubenswrapper[4956]: I1126 16:54:20.996178 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:20 crc kubenswrapper[4956]: E1126 16:54:20.996537 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:22 crc kubenswrapper[4956]: I1126 16:54:22.994370 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:22 crc kubenswrapper[4956]: I1126 16:54:22.994479 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:22 crc kubenswrapper[4956]: I1126 16:54:22.994570 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:22 crc kubenswrapper[4956]: E1126 16:54:22.994789 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:22 crc kubenswrapper[4956]: I1126 16:54:22.994835 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:22 crc kubenswrapper[4956]: E1126 16:54:22.995056 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:22 crc kubenswrapper[4956]: E1126 16:54:22.995437 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:22 crc kubenswrapper[4956]: E1126 16:54:22.995536 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:24 crc kubenswrapper[4956]: I1126 16:54:24.994685 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:24 crc kubenswrapper[4956]: I1126 16:54:24.994772 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:24 crc kubenswrapper[4956]: I1126 16:54:24.996344 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:24 crc kubenswrapper[4956]: E1126 16:54:24.999461 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:24 crc kubenswrapper[4956]: I1126 16:54:24.999904 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:25 crc kubenswrapper[4956]: E1126 16:54:25.000072 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:25 crc kubenswrapper[4956]: E1126 16:54:25.000441 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:25 crc kubenswrapper[4956]: E1126 16:54:25.000900 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:26 crc kubenswrapper[4956]: I1126 16:54:26.994747 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:26 crc kubenswrapper[4956]: I1126 16:54:26.995001 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:26 crc kubenswrapper[4956]: I1126 16:54:26.995150 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:26 crc kubenswrapper[4956]: E1126 16:54:26.995353 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:26 crc kubenswrapper[4956]: E1126 16:54:26.995490 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:26 crc kubenswrapper[4956]: I1126 16:54:26.995452 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:26 crc kubenswrapper[4956]: E1126 16:54:26.996334 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:26 crc kubenswrapper[4956]: E1126 16:54:26.996590 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:28 crc kubenswrapper[4956]: I1126 16:54:28.994517 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:28 crc kubenswrapper[4956]: I1126 16:54:28.994553 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:28 crc kubenswrapper[4956]: I1126 16:54:28.994707 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:28 crc kubenswrapper[4956]: I1126 16:54:28.994623 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:28 crc kubenswrapper[4956]: E1126 16:54:28.994933 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:28 crc kubenswrapper[4956]: E1126 16:54:28.995062 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:28 crc kubenswrapper[4956]: E1126 16:54:28.995335 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:28 crc kubenswrapper[4956]: E1126 16:54:28.995408 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:29 crc kubenswrapper[4956]: I1126 16:54:29.996957 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 16:54:29 crc kubenswrapper[4956]: E1126 16:54:29.997252 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:54:30 crc kubenswrapper[4956]: I1126 16:54:30.995045 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:30 crc kubenswrapper[4956]: I1126 16:54:30.995137 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:30 crc kubenswrapper[4956]: I1126 16:54:30.995316 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:30 crc kubenswrapper[4956]: I1126 16:54:30.995091 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:30 crc kubenswrapper[4956]: E1126 16:54:30.995307 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:30 crc kubenswrapper[4956]: E1126 16:54:30.995427 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:30 crc kubenswrapper[4956]: E1126 16:54:30.995509 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:30 crc kubenswrapper[4956]: E1126 16:54:30.995550 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:32 crc kubenswrapper[4956]: I1126 16:54:32.994996 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:32 crc kubenswrapper[4956]: I1126 16:54:32.995074 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:32 crc kubenswrapper[4956]: E1126 16:54:32.995259 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:32 crc kubenswrapper[4956]: E1126 16:54:32.995717 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:32 crc kubenswrapper[4956]: I1126 16:54:32.995738 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:32 crc kubenswrapper[4956]: I1126 16:54:32.995784 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:32 crc kubenswrapper[4956]: E1126 16:54:32.995940 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:32 crc kubenswrapper[4956]: E1126 16:54:32.996090 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:34 crc kubenswrapper[4956]: I1126 16:54:34.995229 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:34 crc kubenswrapper[4956]: I1126 16:54:34.995303 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:34 crc kubenswrapper[4956]: E1126 16:54:34.995071 4956 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Nov 26 16:54:34 crc kubenswrapper[4956]: E1126 16:54:34.998520 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:34 crc kubenswrapper[4956]: I1126 16:54:34.998566 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:34 crc kubenswrapper[4956]: I1126 16:54:34.998610 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:34 crc kubenswrapper[4956]: E1126 16:54:34.998809 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:34 crc kubenswrapper[4956]: E1126 16:54:34.999199 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:34 crc kubenswrapper[4956]: E1126 16:54:34.999602 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:35 crc kubenswrapper[4956]: E1126 16:54:35.128266 4956 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.759313 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/1.log" Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.760439 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/0.log" Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.760529 4956 generic.go:334] "Generic (PLEG): container finished" podID="80ac0050-7f55-4845-ace7-a89f2442f884" containerID="bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb" exitCode=1 Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.760586 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerDied","Data":"bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb"} Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.760663 4956 scope.go:117] "RemoveContainer" containerID="7777d99572a1af61d5a36f5b0faf2b3f359bccd7e726dedd33e795e25b1e2cb4" Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.761424 4956 scope.go:117] "RemoveContainer" containerID="bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb" Nov 26 16:54:35 crc kubenswrapper[4956]: E1126 16:54:35.761912 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-m4dm4_openshift-multus(80ac0050-7f55-4845-ace7-a89f2442f884)\"" pod="openshift-multus/multus-m4dm4" podUID="80ac0050-7f55-4845-ace7-a89f2442f884" Nov 26 16:54:35 crc kubenswrapper[4956]: I1126 16:54:35.794919 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=15.794893277 podStartE2EDuration="15.794893277s" podCreationTimestamp="2025-11-26 16:54:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:25.038129359 +0000 UTC m=+110.734089981" watchObservedRunningTime="2025-11-26 16:54:35.794893277 +0000 UTC m=+121.490853859" Nov 26 16:54:36 crc kubenswrapper[4956]: I1126 16:54:36.766921 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/1.log" Nov 26 16:54:36 crc kubenswrapper[4956]: I1126 16:54:36.994506 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:36 crc kubenswrapper[4956]: I1126 16:54:36.994553 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:36 crc kubenswrapper[4956]: I1126 16:54:36.994763 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:36 crc kubenswrapper[4956]: E1126 16:54:36.994791 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:36 crc kubenswrapper[4956]: E1126 16:54:36.994937 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:36 crc kubenswrapper[4956]: I1126 16:54:36.994982 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:36 crc kubenswrapper[4956]: E1126 16:54:36.995046 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:36 crc kubenswrapper[4956]: E1126 16:54:36.995845 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:38 crc kubenswrapper[4956]: I1126 16:54:38.994501 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:38 crc kubenswrapper[4956]: I1126 16:54:38.994554 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:38 crc kubenswrapper[4956]: I1126 16:54:38.994698 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:38 crc kubenswrapper[4956]: I1126 16:54:38.994832 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:38 crc kubenswrapper[4956]: E1126 16:54:38.994837 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:38 crc kubenswrapper[4956]: E1126 16:54:38.995114 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:38 crc kubenswrapper[4956]: E1126 16:54:38.995169 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:38 crc kubenswrapper[4956]: E1126 16:54:38.995256 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:40 crc kubenswrapper[4956]: E1126 16:54:40.129561 4956 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 16:54:40 crc kubenswrapper[4956]: I1126 16:54:40.995004 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:40 crc kubenswrapper[4956]: I1126 16:54:40.995107 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:40 crc kubenswrapper[4956]: I1126 16:54:40.995024 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:40 crc kubenswrapper[4956]: E1126 16:54:40.995179 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:40 crc kubenswrapper[4956]: I1126 16:54:40.995025 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:40 crc kubenswrapper[4956]: E1126 16:54:40.995311 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:40 crc kubenswrapper[4956]: E1126 16:54:40.995911 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:40 crc kubenswrapper[4956]: E1126 16:54:40.995759 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:42 crc kubenswrapper[4956]: I1126 16:54:42.995242 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:42 crc kubenswrapper[4956]: I1126 16:54:42.995317 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:42 crc kubenswrapper[4956]: E1126 16:54:42.995930 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:42 crc kubenswrapper[4956]: I1126 16:54:42.995430 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:42 crc kubenswrapper[4956]: I1126 16:54:42.995317 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:42 crc kubenswrapper[4956]: E1126 16:54:42.996063 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:42 crc kubenswrapper[4956]: E1126 16:54:42.996197 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:42 crc kubenswrapper[4956]: E1126 16:54:42.996332 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:43 crc kubenswrapper[4956]: I1126 16:54:43.996104 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 16:54:43 crc kubenswrapper[4956]: E1126 16:54:43.996364 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f2thc_openshift-ovn-kubernetes(46bd4702-b565-4452-be2c-36a3f1f48621)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" Nov 26 16:54:44 crc kubenswrapper[4956]: I1126 16:54:44.995242 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:44 crc kubenswrapper[4956]: I1126 16:54:44.995285 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:44 crc kubenswrapper[4956]: I1126 16:54:44.995272 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:44 crc kubenswrapper[4956]: I1126 16:54:44.995366 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:44 crc kubenswrapper[4956]: E1126 16:54:44.997307 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:44 crc kubenswrapper[4956]: E1126 16:54:44.997440 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:44 crc kubenswrapper[4956]: E1126 16:54:44.997577 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:44 crc kubenswrapper[4956]: E1126 16:54:44.997967 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:45 crc kubenswrapper[4956]: E1126 16:54:45.131344 4956 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 16:54:45 crc kubenswrapper[4956]: I1126 16:54:45.995519 4956 scope.go:117] "RemoveContainer" containerID="bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb" Nov 26 16:54:46 crc kubenswrapper[4956]: I1126 16:54:46.806365 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/1.log" Nov 26 16:54:46 crc kubenswrapper[4956]: I1126 16:54:46.806465 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerStarted","Data":"43e4c0b34591d5c93dae72d209597bd85a17d5098e385aafdcdf66e667e47b4a"} Nov 26 16:54:46 crc kubenswrapper[4956]: I1126 16:54:46.995062 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:46 crc kubenswrapper[4956]: I1126 16:54:46.995133 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:46 crc kubenswrapper[4956]: I1126 16:54:46.995133 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:46 crc kubenswrapper[4956]: I1126 16:54:46.995163 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:46 crc kubenswrapper[4956]: E1126 16:54:46.995277 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:46 crc kubenswrapper[4956]: E1126 16:54:46.995407 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:46 crc kubenswrapper[4956]: E1126 16:54:46.995464 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:46 crc kubenswrapper[4956]: E1126 16:54:46.995529 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:48 crc kubenswrapper[4956]: I1126 16:54:48.994770 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:48 crc kubenswrapper[4956]: I1126 16:54:48.994801 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:48 crc kubenswrapper[4956]: E1126 16:54:48.994965 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:48 crc kubenswrapper[4956]: I1126 16:54:48.994965 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:48 crc kubenswrapper[4956]: I1126 16:54:48.995099 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:48 crc kubenswrapper[4956]: E1126 16:54:48.995117 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:48 crc kubenswrapper[4956]: E1126 16:54:48.995320 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:48 crc kubenswrapper[4956]: E1126 16:54:48.995368 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:50 crc kubenswrapper[4956]: E1126 16:54:50.132766 4956 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 16:54:50 crc kubenswrapper[4956]: I1126 16:54:50.995149 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:50 crc kubenswrapper[4956]: I1126 16:54:50.995225 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:50 crc kubenswrapper[4956]: E1126 16:54:50.995303 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:50 crc kubenswrapper[4956]: I1126 16:54:50.995250 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:50 crc kubenswrapper[4956]: I1126 16:54:50.995498 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:50 crc kubenswrapper[4956]: E1126 16:54:50.995529 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:50 crc kubenswrapper[4956]: E1126 16:54:50.995678 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:50 crc kubenswrapper[4956]: E1126 16:54:50.995713 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:52 crc kubenswrapper[4956]: I1126 16:54:52.995112 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:52 crc kubenswrapper[4956]: I1126 16:54:52.995185 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:52 crc kubenswrapper[4956]: I1126 16:54:52.995224 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:52 crc kubenswrapper[4956]: E1126 16:54:52.995423 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:52 crc kubenswrapper[4956]: I1126 16:54:52.995509 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:52 crc kubenswrapper[4956]: E1126 16:54:52.995676 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:52 crc kubenswrapper[4956]: E1126 16:54:52.995917 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:52 crc kubenswrapper[4956]: E1126 16:54:52.995975 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:54 crc kubenswrapper[4956]: I1126 16:54:54.994374 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:54 crc kubenswrapper[4956]: I1126 16:54:54.994374 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:54 crc kubenswrapper[4956]: I1126 16:54:54.994516 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:54 crc kubenswrapper[4956]: I1126 16:54:54.994585 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:54 crc kubenswrapper[4956]: E1126 16:54:54.996347 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:54 crc kubenswrapper[4956]: E1126 16:54:54.998492 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:54 crc kubenswrapper[4956]: E1126 16:54:54.998844 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:54 crc kubenswrapper[4956]: E1126 16:54:54.999151 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:54 crc kubenswrapper[4956]: I1126 16:54:54.999900 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 16:54:55 crc kubenswrapper[4956]: E1126 16:54:55.139374 4956 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Nov 26 16:54:55 crc kubenswrapper[4956]: I1126 16:54:55.965740 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b8nt2"] Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.000599 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/3.log" Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.004244 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerStarted","Data":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.004272 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:56 crc kubenswrapper[4956]: E1126 16:54:56.004480 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.005073 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.048598 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podStartSLOduration=116.048568332 podStartE2EDuration="1m56.048568332s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:54:56.048256073 +0000 UTC m=+141.744216635" watchObservedRunningTime="2025-11-26 16:54:56.048568332 +0000 UTC m=+141.744528924" Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.996273 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:56 crc kubenswrapper[4956]: E1126 16:54:56.996534 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.997033 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:56 crc kubenswrapper[4956]: I1126 16:54:56.997165 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:56 crc kubenswrapper[4956]: E1126 16:54:56.997200 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:56 crc kubenswrapper[4956]: E1126 16:54:56.997386 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:57 crc kubenswrapper[4956]: I1126 16:54:57.995069 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:57 crc kubenswrapper[4956]: E1126 16:54:57.995403 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:54:58 crc kubenswrapper[4956]: I1126 16:54:58.995131 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:54:58 crc kubenswrapper[4956]: I1126 16:54:58.995145 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:54:58 crc kubenswrapper[4956]: E1126 16:54:58.995835 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Nov 26 16:54:58 crc kubenswrapper[4956]: I1126 16:54:58.995156 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:54:58 crc kubenswrapper[4956]: E1126 16:54:58.996034 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Nov 26 16:54:58 crc kubenswrapper[4956]: E1126 16:54:58.996154 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Nov 26 16:54:59 crc kubenswrapper[4956]: I1126 16:54:59.995235 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:54:59 crc kubenswrapper[4956]: E1126 16:54:59.995483 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-b8nt2" podUID="0956e12e-96b8-4b83-bc0a-666639b36586" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.994628 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.994728 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.994842 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.998107 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.998390 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.998506 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 26 16:55:00 crc kubenswrapper[4956]: I1126 16:55:00.999086 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 26 16:55:01 crc kubenswrapper[4956]: I1126 16:55:01.995255 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:55:01 crc kubenswrapper[4956]: I1126 16:55:01.998142 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 26 16:55:01 crc kubenswrapper[4956]: I1126 16:55:01.999454 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.858172 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:02 crc kubenswrapper[4956]: E1126 16:55:02.858448 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:57:04.858402411 +0000 UTC m=+270.554362993 (durationBeforeRetry 2m2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.858908 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.859208 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.859425 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.859696 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.861592 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.869431 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.869725 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:55:02 crc kubenswrapper[4956]: I1126 16:55:02.871960 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.128545 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.144230 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.155438 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:55:03 crc kubenswrapper[4956]: W1126 16:55:03.397539 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9d751cbb_f2e2_430d_9754_c882a5e924a5.slice/crio-a8faa8c76d613bffdbf6b253071fe6f9c46cc04e321e29b00eae1420fe896cd2 WatchSource:0}: Error finding container a8faa8c76d613bffdbf6b253071fe6f9c46cc04e321e29b00eae1420fe896cd2: Status 404 returned error can't find the container with id a8faa8c76d613bffdbf6b253071fe6f9c46cc04e321e29b00eae1420fe896cd2 Nov 26 16:55:03 crc kubenswrapper[4956]: W1126 16:55:03.409786 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-8bde84f4657386dcb950e26aacc0f5d630fd03733f8b3347f31f33aca92becff WatchSource:0}: Error finding container 8bde84f4657386dcb950e26aacc0f5d630fd03733f8b3347f31f33aca92becff: Status 404 returned error can't find the container with id 8bde84f4657386dcb950e26aacc0f5d630fd03733f8b3347f31f33aca92becff Nov 26 16:55:03 crc kubenswrapper[4956]: W1126 16:55:03.443592 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b6479f0_333b_4a96_9adf_2099afdc2447.slice/crio-6af66ece94e8279a4abfb81c0ad4d2056b6363c160fa9b81a02f57c0f1613216 WatchSource:0}: Error finding container 6af66ece94e8279a4abfb81c0ad4d2056b6363c160fa9b81a02f57c0f1613216: Status 404 returned error can't find the container with id 6af66ece94e8279a4abfb81c0ad4d2056b6363c160fa9b81a02f57c0f1613216 Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.846657 4956 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.898768 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8tmqv"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.899516 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.903034 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v5xlx"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.904224 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.904445 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.905651 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.905653 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.906701 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.907266 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-ssdwk"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.907807 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.908542 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.909523 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.909704 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.910032 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.910123 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.910848 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.911723 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.912644 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.914402 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7vqnn"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.920422 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fx7df"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.922537 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.925469 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.942207 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.942434 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.942574 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.942597 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.943416 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.943535 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.943754 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.944148 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.944416 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.944787 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.944826 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945017 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945082 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945017 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945352 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945378 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945645 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945691 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945802 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945885 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.945916 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.946050 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.946598 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.947243 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.948785 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.949046 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.949202 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.949349 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.949475 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.949586 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.951706 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.951889 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.951981 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952037 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952061 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952161 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952182 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952301 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952313 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952434 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952519 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952563 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952669 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952733 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952777 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952926 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952962 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.953011 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952305 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.953172 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.953947 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954105 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954198 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.952530 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954310 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954410 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954436 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954530 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954597 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954636 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.954768 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dl28z"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.955587 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nd5x4"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.962045 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.963280 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.963915 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j2x66"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.964323 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.964346 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.964409 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.967160 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-cwlp6"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.967773 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.983636 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq"] Nov 26 16:55:03 crc kubenswrapper[4956]: I1126 16:55:03.986792 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.029973 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.030243 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.041239 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"87c0431023b00c1f4cb72080999a8a3b76a98a96f15231a3157772c984045d42"} Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.041315 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a8faa8c76d613bffdbf6b253071fe6f9c46cc04e321e29b00eae1420fe896cd2"} Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.045853 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"ee9272d1479415de8e17cd0f7c99b8cb509512ab76fcaf236a46f13775ae36fb"} Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.045946 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"6af66ece94e8279a4abfb81c0ad4d2056b6363c160fa9b81a02f57c0f1613216"} Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.046597 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.059163 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"c3cd21ea39cea6b51a5e4ac38e41ba16d6fa4033bb1e93e2a7fd0470f3bd90ba"} Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.059230 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"8bde84f4657386dcb950e26aacc0f5d630fd03733f8b3347f31f33aca92becff"} Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.067693 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.068001 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077415 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q57kt\" (UniqueName: \"kubernetes.io/projected/fa92cf8a-56c1-4b38-889c-23a578b74b90-kube-api-access-q57kt\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077458 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-serving-cert\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077486 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077509 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077526 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be8528e4-668a-48fb-b8b9-556f9d57118a-trusted-ca\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077545 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-policies\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077563 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077583 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pbdzf\" (UniqueName: \"kubernetes.io/projected/166d8e6f-9af9-4212-9593-0b47d27918d1-kube-api-access-pbdzf\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077599 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-config\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077622 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-etcd-client\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077639 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa92cf8a-56c1-4b38-889c-23a578b74b90-serving-cert\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077656 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pqcqp\" (UniqueName: \"kubernetes.io/projected/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-kube-api-access-pqcqp\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077679 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3f11c7e-1585-49d4-8675-2b4234fd005e-auth-proxy-config\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077697 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077714 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-console-config\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077732 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-image-import-ca\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077749 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077766 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1fae2fb-ef9b-4085-b673-7514924aa516-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077785 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bea9e494-4444-40f7-8b17-f5ee6293d553-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077804 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdfsf\" (UniqueName: \"kubernetes.io/projected/5e23939f-1541-494e-af81-1cf62ff004b0-kube-api-access-sdfsf\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077820 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-audit\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077836 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-encryption-config\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077853 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-config\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077894 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vxsrf\" (UniqueName: \"kubernetes.io/projected/b9b65044-9326-4a72-a933-e84012a29211-kube-api-access-vxsrf\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077914 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/183ca9fa-6509-45fb-8f0a-41161814ecbb-node-pullsecrets\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077933 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-km9l6\" (UniqueName: \"kubernetes.io/projected/996f361a-f4aa-4d79-b12c-3e357f126f43-kube-api-access-km9l6\") pod \"downloads-7954f5f757-cwlp6\" (UID: \"996f361a-f4aa-4d79-b12c-3e357f126f43\") " pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077951 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-config\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077968 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-dir\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.077985 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96kd8\" (UniqueName: \"kubernetes.io/projected/bea9e494-4444-40f7-8b17-f5ee6293d553-kube-api-access-96kd8\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078005 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-encryption-config\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078025 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n7tq6\" (UniqueName: \"kubernetes.io/projected/f3f11c7e-1585-49d4-8675-2b4234fd005e-kube-api-access-n7tq6\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078045 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078066 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-service-ca\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078124 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e23939f-1541-494e-af81-1cf62ff004b0-audit-dir\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078166 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be8528e4-668a-48fb-b8b9-556f9d57118a-config\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078195 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qnjqh\" (UniqueName: \"kubernetes.io/projected/6aa20dbb-4a6d-4a96-bf62-7c4e337285ed-kube-api-access-qnjqh\") pod \"cluster-samples-operator-665b6dd947-z5f2l\" (UID: \"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078302 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078334 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8lgn\" (UniqueName: \"kubernetes.io/projected/5552a148-c82e-4afb-9a27-9ce40714bdb0-kube-api-access-l8lgn\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078509 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078535 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078558 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bea9e494-4444-40f7-8b17-f5ee6293d553-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078583 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s9kx7\" (UniqueName: \"kubernetes.io/projected/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-kube-api-access-s9kx7\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078626 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f3f11c7e-1585-49d4-8675-2b4234fd005e-machine-approver-tls\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078712 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bwtsz\" (UniqueName: \"kubernetes.io/projected/b1fae2fb-ef9b-4085-b673-7514924aa516-kube-api-access-bwtsz\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078787 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-etcd-serving-ca\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.078821 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.079029 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.079099 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-serving-cert\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.079125 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/166d8e6f-9af9-4212-9593-0b47d27918d1-console-serving-cert\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.079205 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2f92k\" (UniqueName: \"kubernetes.io/projected/be8528e4-668a-48fb-b8b9-556f9d57118a-kube-api-access-2f92k\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.079232 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-config\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.080854 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5htxr\" (UniqueName: \"kubernetes.io/projected/724bfa7f-010c-4298-a11a-50ea0f5fe198-kube-api-access-5htxr\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.080925 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.080993 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-etcd-client\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081056 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-service-ca\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081103 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/724bfa7f-010c-4298-a11a-50ea0f5fe198-config\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081132 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-audit-policies\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081153 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-client-ca\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081202 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1fae2fb-ef9b-4085-b673-7514924aa516-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081435 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081472 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-client-ca\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081495 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/724bfa7f-010c-4298-a11a-50ea0f5fe198-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081637 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081678 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081704 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/166d8e6f-9af9-4212-9593-0b47d27918d1-console-oauth-config\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081729 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5pkq2\" (UniqueName: \"kubernetes.io/projected/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-kube-api-access-5pkq2\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081747 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2jxwp\" (UniqueName: \"kubernetes.io/projected/183ca9fa-6509-45fb-8f0a-41161814ecbb-kube-api-access-2jxwp\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081770 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b65044-9326-4a72-a933-e84012a29211-serving-cert\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081823 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be8528e4-668a-48fb-b8b9-556f9d57118a-serving-cert\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081883 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-config\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081914 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/724bfa7f-010c-4298-a11a-50ea0f5fe198-images\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081949 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-ca\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.081982 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-trusted-ca-bundle\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.082015 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-serving-cert\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.082412 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.082436 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-serving-cert\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.082565 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.082802 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.083007 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.084664 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.084914 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085139 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085350 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085448 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085499 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3f11c7e-1585-49d4-8675-2b4234fd005e-config\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085556 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-client\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085593 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085614 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085623 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-oauth-serving-cert\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085652 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6aa20dbb-4a6d-4a96-bf62-7c4e337285ed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z5f2l\" (UID: \"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085697 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/183ca9fa-6509-45fb-8f0a-41161814ecbb-audit-dir\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.085721 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.086088 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.086259 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.086603 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.088718 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.100944 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.102765 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.115628 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.121064 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.121208 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.130567 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.130814 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131117 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131164 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131503 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131670 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131780 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131824 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.131917 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.132204 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.132327 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.132775 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.135597 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.135732 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.135819 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.136000 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.136107 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.138639 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.140088 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.140203 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.140333 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.140097 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.143565 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.143636 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v5xlx"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.143932 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.144023 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7vqnn"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.146220 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xjxmw"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.146906 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.149534 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.149578 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.150706 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.151237 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.153907 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.154162 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.155160 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9wvt8"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.155306 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.155986 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.156467 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.156534 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.156798 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.156469 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.157724 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ssdwk"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.157791 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.158970 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.159720 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.159932 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.160555 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.160986 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-r245h"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.163979 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.164110 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.166174 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.173116 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.173657 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.174151 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.175828 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j2x66"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.175882 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwzxf"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.175926 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.179823 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.181517 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.193166 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.194118 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211465 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pqcqp\" (UniqueName: \"kubernetes.io/projected/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-kube-api-access-pqcqp\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211534 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3f11c7e-1585-49d4-8675-2b4234fd005e-auth-proxy-config\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211582 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-console-config\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211618 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211647 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-image-import-ca\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211672 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211700 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1fae2fb-ef9b-4085-b673-7514924aa516-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211735 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bea9e494-4444-40f7-8b17-f5ee6293d553-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211766 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdfsf\" (UniqueName: \"kubernetes.io/projected/5e23939f-1541-494e-af81-1cf62ff004b0-kube-api-access-sdfsf\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211795 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-audit\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211821 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-encryption-config\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211853 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2ec484-4014-4235-bbe5-aa47e9f0f083-config\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211899 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9a53de27-4165-434e-8f23-6180d37c530c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211933 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vxsrf\" (UniqueName: \"kubernetes.io/projected/b9b65044-9326-4a72-a933-e84012a29211-kube-api-access-vxsrf\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.211957 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/706f4561-843f-4b61-bede-4d4c02d8dad6-images\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212001 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/183ca9fa-6509-45fb-8f0a-41161814ecbb-node-pullsecrets\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212027 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-config\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212051 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-config\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212078 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-dir\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212105 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-km9l6\" (UniqueName: \"kubernetes.io/projected/996f361a-f4aa-4d79-b12c-3e357f126f43-kube-api-access-km9l6\") pod \"downloads-7954f5f757-cwlp6\" (UID: \"996f361a-f4aa-4d79-b12c-3e357f126f43\") " pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212129 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-encryption-config\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212154 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n7tq6\" (UniqueName: \"kubernetes.io/projected/f3f11c7e-1585-49d4-8675-2b4234fd005e-kube-api-access-n7tq6\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212182 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce58bfc7-f777-4db0-8690-ce9a4137a92c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212213 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212238 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96kd8\" (UniqueName: \"kubernetes.io/projected/bea9e494-4444-40f7-8b17-f5ee6293d553-kube-api-access-96kd8\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212264 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e23939f-1541-494e-af81-1cf62ff004b0-audit-dir\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212287 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be8528e4-668a-48fb-b8b9-556f9d57118a-config\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212313 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qnjqh\" (UniqueName: \"kubernetes.io/projected/6aa20dbb-4a6d-4a96-bf62-7c4e337285ed-kube-api-access-qnjqh\") pod \"cluster-samples-operator-665b6dd947-z5f2l\" (UID: \"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212342 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/229601a4-1c80-4ed2-9330-4297573d9af6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212370 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212398 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8lgn\" (UniqueName: \"kubernetes.io/projected/5552a148-c82e-4afb-9a27-9ce40714bdb0-kube-api-access-l8lgn\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212421 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-service-ca\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212449 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212470 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212491 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bea9e494-4444-40f7-8b17-f5ee6293d553-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212536 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s9kx7\" (UniqueName: \"kubernetes.io/projected/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-kube-api-access-s9kx7\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212566 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f3f11c7e-1585-49d4-8675-2b4234fd005e-machine-approver-tls\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212575 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212592 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229601a4-1c80-4ed2-9330-4297573d9af6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212891 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae2ec484-4014-4235-bbe5-aa47e9f0f083-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212935 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bwtsz\" (UniqueName: \"kubernetes.io/projected/b1fae2fb-ef9b-4085-b673-7514924aa516-kube-api-access-bwtsz\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212959 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/229601a4-1c80-4ed2-9330-4297573d9af6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.212986 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-etcd-serving-ca\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213007 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213025 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1761b713-5b59-4758-99cc-11a2e8eff4ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213059 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-serving-cert\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213077 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/166d8e6f-9af9-4212-9593-0b47d27918d1-console-serving-cert\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213094 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2f92k\" (UniqueName: \"kubernetes.io/projected/be8528e4-668a-48fb-b8b9-556f9d57118a-kube-api-access-2f92k\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213112 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-config\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213135 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213153 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5htxr\" (UniqueName: \"kubernetes.io/projected/724bfa7f-010c-4298-a11a-50ea0f5fe198-kube-api-access-5htxr\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213168 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-etcd-client\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213186 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-service-ca\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213206 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/724bfa7f-010c-4298-a11a-50ea0f5fe198-config\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213233 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213253 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-audit-policies\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213268 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-client-ca\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213288 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1fae2fb-ef9b-4085-b673-7514924aa516-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213311 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-client-ca\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213367 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/724bfa7f-010c-4298-a11a-50ea0f5fe198-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213391 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmswz\" (UniqueName: \"kubernetes.io/projected/1761b713-5b59-4758-99cc-11a2e8eff4ae-kube-api-access-wmswz\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213411 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/706f4561-843f-4b61-bede-4d4c02d8dad6-proxy-tls\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213458 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213477 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/706f4561-843f-4b61-bede-4d4c02d8dad6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213498 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213516 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213535 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/166d8e6f-9af9-4212-9593-0b47d27918d1-console-oauth-config\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213559 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wm65x\" (UniqueName: \"kubernetes.io/projected/706f4561-843f-4b61-bede-4d4c02d8dad6-kube-api-access-wm65x\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213586 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5pkq2\" (UniqueName: \"kubernetes.io/projected/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-kube-api-access-5pkq2\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213604 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ce58bfc7-f777-4db0-8690-ce9a4137a92c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213631 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2jxwp\" (UniqueName: \"kubernetes.io/projected/183ca9fa-6509-45fb-8f0a-41161814ecbb-kube-api-access-2jxwp\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213649 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b65044-9326-4a72-a933-e84012a29211-serving-cert\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213667 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcft7\" (UniqueName: \"kubernetes.io/projected/ce58bfc7-f777-4db0-8690-ce9a4137a92c-kube-api-access-qcft7\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213691 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-config\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213710 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/724bfa7f-010c-4298-a11a-50ea0f5fe198-images\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213728 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1761b713-5b59-4758-99cc-11a2e8eff4ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213752 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-ca\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213772 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be8528e4-668a-48fb-b8b9-556f9d57118a-serving-cert\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213792 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-trusted-ca-bundle\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213810 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4nbhl\" (UniqueName: \"kubernetes.io/projected/9a53de27-4165-434e-8f23-6180d37c530c-kube-api-access-4nbhl\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213830 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213848 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-serving-cert\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213879 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213899 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3f11c7e-1585-49d4-8675-2b4234fd005e-config\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213919 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-client\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213942 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-serving-cert\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213963 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9a53de27-4165-434e-8f23-6180d37c530c-proxy-tls\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.213984 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214001 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-oauth-serving-cert\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214021 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6aa20dbb-4a6d-4a96-bf62-7c4e337285ed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z5f2l\" (UID: \"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214042 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214063 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce58bfc7-f777-4db0-8690-ce9a4137a92c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214086 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/183ca9fa-6509-45fb-8f0a-41161814ecbb-audit-dir\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214107 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-serving-cert\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214127 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214149 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q57kt\" (UniqueName: \"kubernetes.io/projected/fa92cf8a-56c1-4b38-889c-23a578b74b90-kube-api-access-q57kt\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214168 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214186 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be8528e4-668a-48fb-b8b9-556f9d57118a-trusted-ca\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214203 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae2ec484-4014-4235-bbe5-aa47e9f0f083-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214229 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-policies\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214248 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214265 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pbdzf\" (UniqueName: \"kubernetes.io/projected/166d8e6f-9af9-4212-9593-0b47d27918d1-kube-api-access-pbdzf\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214285 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-config\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214308 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-etcd-client\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.214327 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa92cf8a-56c1-4b38-889c-23a578b74b90-serving-cert\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.215721 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-etcd-serving-ca\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.216036 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/f3f11c7e-1585-49d4-8675-2b4234fd005e-auth-proxy-config\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.216169 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.216998 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-console-config\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.217136 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-oauth-serving-cert\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.217737 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.218664 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-image-import-ca\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.219997 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.221429 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-trusted-ca-bundle\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.223229 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-serving-cert\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.223229 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-ca\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.223928 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/724bfa7f-010c-4298-a11a-50ea0f5fe198-images\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.223939 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b65044-9326-4a72-a933-e84012a29211-serving-cert\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.224408 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.224788 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-client-ca\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.225601 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.226216 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-config\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.226397 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.226763 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bea9e494-4444-40f7-8b17-f5ee6293d553-available-featuregates\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.226781 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-client-ca\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.226840 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b1fae2fb-ef9b-4085-b673-7514924aa516-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.228430 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/183ca9fa-6509-45fb-8f0a-41161814ecbb-node-pullsecrets\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.231395 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-dir\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.232616 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/183ca9fa-6509-45fb-8f0a-41161814ecbb-audit-dir\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.234417 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-service-ca\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.234774 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-etcd-client\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.235031 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-audit\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.235538 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/724bfa7f-010c-4298-a11a-50ea0f5fe198-config\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.236297 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/be8528e4-668a-48fb-b8b9-556f9d57118a-config\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.236338 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5e23939f-1541-494e-af81-1cf62ff004b0-audit-dir\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.236336 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/166d8e6f-9af9-4212-9593-0b47d27918d1-console-serving-cert\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.237121 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.237531 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-shfrr"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.237671 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/724bfa7f-010c-4298-a11a-50ea0f5fe198-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.238364 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.239619 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.240920 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.243008 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.243381 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b1fae2fb-ef9b-4085-b673-7514924aa516-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.243666 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-service-ca-bundle\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.244001 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-config\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.244272 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bea9e494-4444-40f7-8b17-f5ee6293d553-serving-cert\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.245019 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/166d8e6f-9af9-4212-9593-0b47d27918d1-console-oauth-config\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.245399 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/be8528e4-668a-48fb-b8b9-556f9d57118a-serving-cert\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.245664 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-policies\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.245999 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.246698 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-config\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.247022 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.247152 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.247220 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fx7df"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.247305 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-24z97"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.247972 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7txdb"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.248611 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.249034 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.249484 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.249595 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/183ca9fa-6509-45fb-8f0a-41161814ecbb-config\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.249656 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.249614 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.250017 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.250239 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.251226 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.251984 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-cwlp6"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.252062 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.252148 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dl28z"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.252207 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.252290 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-lh9p2"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.252805 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8htxc"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.253315 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.255680 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nd5x4"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.255767 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9wvt8"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.255826 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.255994 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.256358 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.256579 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.256779 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.258029 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.258085 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.259096 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.260438 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.260603 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-serving-cert\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.260705 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa92cf8a-56c1-4b38-889c-23a578b74b90-serving-cert\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.261589 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.261644 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.261782 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.261843 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/166d8e6f-9af9-4212-9593-0b47d27918d1-trusted-ca-bundle\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.262158 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-encryption-config\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.262460 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.262614 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.263193 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.263514 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.263925 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.264460 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.265287 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xjxmw"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.265747 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-xk6gr"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.266580 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.266735 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8tmqv"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.267195 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f3f11c7e-1585-49d4-8675-2b4234fd005e-config\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.267566 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-config\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.267786 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/be8528e4-668a-48fb-b8b9-556f9d57118a-trusted-ca\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.268131 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-config\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.268183 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-audit-policies\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.268209 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-service-ca\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.268234 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-24z97"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.269001 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/5e23939f-1541-494e-af81-1cf62ff004b0-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.269058 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7txdb"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.269236 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.270037 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-vkf2v"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.270238 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/6aa20dbb-4a6d-4a96-bf62-7c4e337285ed-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-z5f2l\" (UID: \"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.273252 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-etcd-client\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.273308 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e23939f-1541-494e-af81-1cf62ff004b0-etcd-client\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.273790 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-serving-cert\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.273929 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-serving-cert\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.273974 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/f3f11c7e-1585-49d4-8675-2b4234fd005e-machine-approver-tls\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.273991 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwzxf"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.274135 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.275124 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/183ca9fa-6509-45fb-8f0a-41161814ecbb-encryption-config\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.278348 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-shfrr"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.280491 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.282656 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.286605 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.290318 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.291619 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xk6gr"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.292779 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.294808 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.296140 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.297646 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.299069 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.300162 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.301383 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vkf2v"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.302237 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.302783 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8htxc"] Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315110 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmswz\" (UniqueName: \"kubernetes.io/projected/1761b713-5b59-4758-99cc-11a2e8eff4ae-kube-api-access-wmswz\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315143 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/706f4561-843f-4b61-bede-4d4c02d8dad6-proxy-tls\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315165 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/706f4561-843f-4b61-bede-4d4c02d8dad6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315190 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wm65x\" (UniqueName: \"kubernetes.io/projected/706f4561-843f-4b61-bede-4d4c02d8dad6-kube-api-access-wm65x\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315226 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ce58bfc7-f777-4db0-8690-ce9a4137a92c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315245 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcft7\" (UniqueName: \"kubernetes.io/projected/ce58bfc7-f777-4db0-8690-ce9a4137a92c-kube-api-access-qcft7\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315278 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1761b713-5b59-4758-99cc-11a2e8eff4ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315299 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4nbhl\" (UniqueName: \"kubernetes.io/projected/9a53de27-4165-434e-8f23-6180d37c530c-kube-api-access-4nbhl\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315320 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9a53de27-4165-434e-8f23-6180d37c530c-proxy-tls\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315346 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce58bfc7-f777-4db0-8690-ce9a4137a92c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315375 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae2ec484-4014-4235-bbe5-aa47e9f0f083-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315432 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2ec484-4014-4235-bbe5-aa47e9f0f083-config\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315454 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9a53de27-4165-434e-8f23-6180d37c530c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315490 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/706f4561-843f-4b61-bede-4d4c02d8dad6-images\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315535 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce58bfc7-f777-4db0-8690-ce9a4137a92c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315572 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/229601a4-1c80-4ed2-9330-4297573d9af6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315609 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229601a4-1c80-4ed2-9330-4297573d9af6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315633 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae2ec484-4014-4235-bbe5-aa47e9f0f083-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315669 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/229601a4-1c80-4ed2-9330-4297573d9af6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.315693 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1761b713-5b59-4758-99cc-11a2e8eff4ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.316299 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/706f4561-843f-4b61-bede-4d4c02d8dad6-auth-proxy-config\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.316588 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/9a53de27-4165-434e-8f23-6180d37c530c-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.316623 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ae2ec484-4014-4235-bbe5-aa47e9f0f083-config\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.317338 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/ce58bfc7-f777-4db0-8690-ce9a4137a92c-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.318312 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ae2ec484-4014-4235-bbe5-aa47e9f0f083-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.319661 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/ce58bfc7-f777-4db0-8690-ce9a4137a92c-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.323501 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.343361 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.362472 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.383288 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.410527 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.422296 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.442111 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.462036 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.484523 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.502718 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.523346 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.543737 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.563017 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.567520 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/229601a4-1c80-4ed2-9330-4297573d9af6-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.583317 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.590581 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/229601a4-1c80-4ed2-9330-4297573d9af6-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.603669 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.622607 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.631304 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/9a53de27-4165-434e-8f23-6180d37c530c-proxy-tls\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.644412 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.662301 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.688417 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.703335 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.723768 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.743708 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.763065 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.783213 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.803137 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.822602 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.843564 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.863738 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.873035 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1761b713-5b59-4758-99cc-11a2e8eff4ae-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.886407 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.903343 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.909260 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1761b713-5b59-4758-99cc-11a2e8eff4ae-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.923628 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.942526 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.963927 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.967331 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/706f4561-843f-4b61-bede-4d4c02d8dad6-images\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:04 crc kubenswrapper[4956]: I1126 16:55:04.983169 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.003346 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.010773 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/706f4561-843f-4b61-bede-4d4c02d8dad6-proxy-tls\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.025263 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.043481 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.062384 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.096291 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.103308 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.143015 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.163626 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.183209 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.220981 4956 request.go:700] Waited for 1.005935186s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-apiserver-operator/serviceaccounts/openshift-apiserver-operator/token Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.237242 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2jxwp\" (UniqueName: \"kubernetes.io/projected/183ca9fa-6509-45fb-8f0a-41161814ecbb-kube-api-access-2jxwp\") pod \"apiserver-76f77b778f-fx7df\" (UID: \"183ca9fa-6509-45fb-8f0a-41161814ecbb\") " pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.255940 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pqcqp\" (UniqueName: \"kubernetes.io/projected/ce09c66b-2af4-47e2-9f7d-bf540aca5b5e-kube-api-access-pqcqp\") pod \"openshift-apiserver-operator-796bbdcf4f-v8mzp\" (UID: \"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.260782 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.263384 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bwtsz\" (UniqueName: \"kubernetes.io/projected/b1fae2fb-ef9b-4085-b673-7514924aa516-kube-api-access-bwtsz\") pod \"openshift-controller-manager-operator-756b6f6bc6-874qk\" (UID: \"b1fae2fb-ef9b-4085-b673-7514924aa516\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.301542 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2f92k\" (UniqueName: \"kubernetes.io/projected/be8528e4-668a-48fb-b8b9-556f9d57118a-kube-api-access-2f92k\") pod \"console-operator-58897d9998-nd5x4\" (UID: \"be8528e4-668a-48fb-b8b9-556f9d57118a\") " pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.304739 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5pkq2\" (UniqueName: \"kubernetes.io/projected/8dd2caa0-c053-4d2b-9b22-dfbf133065ac-kube-api-access-5pkq2\") pod \"authentication-operator-69f744f599-dl28z\" (UID: \"8dd2caa0-c053-4d2b-9b22-dfbf133065ac\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.318649 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdfsf\" (UniqueName: \"kubernetes.io/projected/5e23939f-1541-494e-af81-1cf62ff004b0-kube-api-access-sdfsf\") pod \"apiserver-7bbb656c7d-f2ldn\" (UID: \"5e23939f-1541-494e-af81-1cf62ff004b0\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.344762 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q57kt\" (UniqueName: \"kubernetes.io/projected/fa92cf8a-56c1-4b38-889c-23a578b74b90-kube-api-access-q57kt\") pod \"route-controller-manager-6576b87f9c-kr7n2\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.351307 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.362858 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-km9l6\" (UniqueName: \"kubernetes.io/projected/996f361a-f4aa-4d79-b12c-3e357f126f43-kube-api-access-km9l6\") pod \"downloads-7954f5f757-cwlp6\" (UID: \"996f361a-f4aa-4d79-b12c-3e357f126f43\") " pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.370997 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.381509 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vxsrf\" (UniqueName: \"kubernetes.io/projected/b9b65044-9326-4a72-a933-e84012a29211-kube-api-access-vxsrf\") pod \"controller-manager-879f6c89f-v5xlx\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.407933 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5htxr\" (UniqueName: \"kubernetes.io/projected/724bfa7f-010c-4298-a11a-50ea0f5fe198-kube-api-access-5htxr\") pod \"machine-api-operator-5694c8668f-7vqnn\" (UID: \"724bfa7f-010c-4298-a11a-50ea0f5fe198\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.421589 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s9kx7\" (UniqueName: \"kubernetes.io/projected/fec6a02f-63fc-473e-8a99-fcbe8ce154a6-kube-api-access-s9kx7\") pod \"etcd-operator-b45778765-j2x66\" (UID: \"fec6a02f-63fc-473e-8a99-fcbe8ce154a6\") " pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.440945 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n7tq6\" (UniqueName: \"kubernetes.io/projected/f3f11c7e-1585-49d4-8675-2b4234fd005e-kube-api-access-n7tq6\") pod \"machine-approver-56656f9798-bpv5w\" (UID: \"f3f11c7e-1585-49d4-8675-2b4234fd005e\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.467595 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pbdzf\" (UniqueName: \"kubernetes.io/projected/166d8e6f-9af9-4212-9593-0b47d27918d1-kube-api-access-pbdzf\") pod \"console-f9d7485db-ssdwk\" (UID: \"166d8e6f-9af9-4212-9593-0b47d27918d1\") " pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.470718 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.484436 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.485158 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96kd8\" (UniqueName: \"kubernetes.io/projected/bea9e494-4444-40f7-8b17-f5ee6293d553-kube-api-access-96kd8\") pod \"openshift-config-operator-7777fb866f-zqnvj\" (UID: \"bea9e494-4444-40f7-8b17-f5ee6293d553\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.495687 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.496192 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-fx7df"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.504151 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.507371 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.514507 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8lgn\" (UniqueName: \"kubernetes.io/projected/5552a148-c82e-4afb-9a27-9ce40714bdb0-kube-api-access-l8lgn\") pod \"oauth-openshift-558db77b4-8tmqv\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.524793 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.525488 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.531224 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.538917 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.545142 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.551796 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.563374 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 26 16:55:05 crc kubenswrapper[4956]: W1126 16:55:05.567089 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf3f11c7e_1585_49d4_8675_2b4234fd005e.slice/crio-f92ec567a29d737b0ad1d1f3f89ad978bfa2337957e8856e005297692fa6d628 WatchSource:0}: Error finding container f92ec567a29d737b0ad1d1f3f89ad978bfa2337957e8856e005297692fa6d628: Status 404 returned error can't find the container with id f92ec567a29d737b0ad1d1f3f89ad978bfa2337957e8856e005297692fa6d628 Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.582112 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.584291 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.597887 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.601436 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.610260 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.640480 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qnjqh\" (UniqueName: \"kubernetes.io/projected/6aa20dbb-4a6d-4a96-bf62-7c4e337285ed-kube-api-access-qnjqh\") pod \"cluster-samples-operator-665b6dd947-z5f2l\" (UID: \"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.660889 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.663442 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 26 16:55:05 crc kubenswrapper[4956]: W1126 16:55:05.664517 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podfa92cf8a_56c1_4b38_889c_23a578b74b90.slice/crio-e9dadad485fad8cfb5c1f50a10dc8db68a6177400d6b2892e650b410d69d0de7 WatchSource:0}: Error finding container e9dadad485fad8cfb5c1f50a10dc8db68a6177400d6b2892e650b410d69d0de7: Status 404 returned error can't find the container with id e9dadad485fad8cfb5c1f50a10dc8db68a6177400d6b2892e650b410d69d0de7 Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.682531 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.704677 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.723426 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.735112 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-cwlp6"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.749415 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.758426 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.763416 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.784313 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.799494 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.804382 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.823691 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.844998 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.858086 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-ssdwk"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.862636 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.869652 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" Nov 26 16:55:05 crc kubenswrapper[4956]: W1126 16:55:05.871701 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb1fae2fb_ef9b_4085_b673_7514924aa516.slice/crio-46a26b3d8a1b1492d5a025cb3bfb4a71c9d2c617c61cfb7ebe8ddb720d65e770 WatchSource:0}: Error finding container 46a26b3d8a1b1492d5a025cb3bfb4a71c9d2c617c61cfb7ebe8ddb720d65e770: Status 404 returned error can't find the container with id 46a26b3d8a1b1492d5a025cb3bfb4a71c9d2c617c61cfb7ebe8ddb720d65e770 Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.890003 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.919268 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.920663 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.922054 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.941972 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.957560 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v5xlx"] Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.965510 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 26 16:55:05 crc kubenswrapper[4956]: I1126 16:55:05.984514 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.002195 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.015389 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.023334 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.041274 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-dl28z"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.043653 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.074015 4956 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.080079 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-j2x66"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.082453 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-7vqnn"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.083085 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.104294 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.123035 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" event={"ID":"f3f11c7e-1585-49d4-8675-2b4234fd005e","Type":"ContainerStarted","Data":"f92ec567a29d737b0ad1d1f3f89ad978bfa2337957e8856e005297692fa6d628"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.124323 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.129096 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" event={"ID":"b9b65044-9326-4a72-a933-e84012a29211","Type":"ContainerStarted","Data":"14fc8934430a0279fc8fe39e01d9feaa7897869a5cbc4187b006ec032291d4f8"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.136078 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-cwlp6" event={"ID":"996f361a-f4aa-4d79-b12c-3e357f126f43","Type":"ContainerStarted","Data":"eb99e59873c8803f334ea69dbd8138e03da75391b319742bedeb9163252f0e6e"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.140290 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ssdwk" event={"ID":"166d8e6f-9af9-4212-9593-0b47d27918d1","Type":"ContainerStarted","Data":"c5a86bdc827ac6843887773295eecc7d0c297baeb6ddc08552ec91581426de3e"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.142670 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.143558 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" event={"ID":"fa92cf8a-56c1-4b38-889c-23a578b74b90","Type":"ContainerStarted","Data":"932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.143603 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" event={"ID":"fa92cf8a-56c1-4b38-889c-23a578b74b90","Type":"ContainerStarted","Data":"e9dadad485fad8cfb5c1f50a10dc8db68a6177400d6b2892e650b410d69d0de7"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.143804 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.147509 4956 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kr7n2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.147563 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.151342 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" event={"ID":"b1fae2fb-ef9b-4085-b673-7514924aa516","Type":"ContainerStarted","Data":"46a26b3d8a1b1492d5a025cb3bfb4a71c9d2c617c61cfb7ebe8ddb720d65e770"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.161638 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8tmqv"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.162786 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" event={"ID":"183ca9fa-6509-45fb-8f0a-41161814ecbb","Type":"ContainerStarted","Data":"edaf805f844a7a0c2d0e7ab3747407fc998156c36f1351e04b6da2fb41902969"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.163003 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.169965 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" event={"ID":"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e","Type":"ContainerStarted","Data":"1c50ab843413216eabf20ddf1f931e8555eeeec683c91af6c8bff866bfe1057f"} Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.182281 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.193437 4956 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9b65044_9326_4a72_a933_e84012a29211.slice/crio-14fc8934430a0279fc8fe39e01d9feaa7897869a5cbc4187b006ec032291d4f8\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod183ca9fa_6509_45fb_8f0a_41161814ecbb.slice/crio-conmon-adc2cec84e9d1fd5543b5003cf623e040506475a4f52c630ea372354eddd0b5b.scope\": RecentStats: unable to find data in memory cache]" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.203182 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.220531 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-nd5x4"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.224963 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.240618 4956 request.go:700] Waited for 1.965740845s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-default-metrics-tls&limit=500&resourceVersion=0 Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.243318 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.244928 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.263047 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.283171 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.325718 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmswz\" (UniqueName: \"kubernetes.io/projected/1761b713-5b59-4758-99cc-11a2e8eff4ae-kube-api-access-wmswz\") pod \"kube-storage-version-migrator-operator-b67b599dd-z4gz4\" (UID: \"1761b713-5b59-4758-99cc-11a2e8eff4ae\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.342847 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wm65x\" (UniqueName: \"kubernetes.io/projected/706f4561-843f-4b61-bede-4d4c02d8dad6-kube-api-access-wm65x\") pod \"machine-config-operator-74547568cd-6mkc4\" (UID: \"706f4561-843f-4b61-bede-4d4c02d8dad6\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.346614 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.361816 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/ce58bfc7-f777-4db0-8690-ce9a4137a92c-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.379625 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcft7\" (UniqueName: \"kubernetes.io/projected/ce58bfc7-f777-4db0-8690-ce9a4137a92c-kube-api-access-qcft7\") pod \"cluster-image-registry-operator-dc59b4c8b-jwqpq\" (UID: \"ce58bfc7-f777-4db0-8690-ce9a4137a92c\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.405910 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/229601a4-1c80-4ed2-9330-4297573d9af6-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-tq7kg\" (UID: \"229601a4-1c80-4ed2-9330-4297573d9af6\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.419543 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ae2ec484-4014-4235-bbe5-aa47e9f0f083-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-dbctb\" (UID: \"ae2ec484-4014-4235-bbe5-aa47e9f0f083\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.433780 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.439255 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4nbhl\" (UniqueName: \"kubernetes.io/projected/9a53de27-4165-434e-8f23-6180d37c530c-kube-api-access-4nbhl\") pod \"machine-config-controller-84d6567774-tn8vt\" (UID: \"9a53de27-4165-434e-8f23-6180d37c530c\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.439322 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468116 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8394076a-e17a-4cea-a904-eb269c237f00-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468181 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/949267c7-0029-430b-a356-7059518c503a-metrics-tls\") pod \"dns-operator-744455d44c-9wvt8\" (UID: \"949267c7-0029-430b-a356-7059518c503a\") " pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468259 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8394076a-e17a-4cea-a904-eb269c237f00-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468281 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7srcw\" (UniqueName: \"kubernetes.io/projected/949267c7-0029-430b-a356-7059518c503a-kube-api-access-7srcw\") pod \"dns-operator-744455d44c-9wvt8\" (UID: \"949267c7-0029-430b-a356-7059518c503a\") " pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468321 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-55rk5\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-kube-api-access-55rk5\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468338 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-registry-certificates\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468353 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/00806805-c6d3-4b27-aff6-3a6f6d03a094-metrics-tls\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468407 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-trusted-ca\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468425 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hhcq2\" (UniqueName: \"kubernetes.io/projected/dead528a-47b6-44fc-97bf-17c308cc2033-kube-api-access-hhcq2\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468461 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468478 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00806805-c6d3-4b27-aff6-3a6f6d03a094-trusted-ca\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468513 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468539 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00806805-c6d3-4b27-aff6-3a6f6d03a094-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468584 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468604 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-bound-sa-token\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468622 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-stats-auth\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468655 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rl7fm\" (UniqueName: \"kubernetes.io/projected/00806805-c6d3-4b27-aff6-3a6f6d03a094-kube-api-access-rl7fm\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468674 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-metrics-certs\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468724 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-registry-tls\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468743 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-service-ca-bundle\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468772 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7wzmc\" (UniqueName: \"kubernetes.io/projected/68745ad4-99d4-4028-95fb-0d70e3fba519-kube-api-access-7wzmc\") pod \"migrator-59844c95c7-sb68p\" (UID: \"68745ad4-99d4-4028-95fb-0d70e3fba519\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468800 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-default-certificate\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.468817 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64qlk\" (UniqueName: \"kubernetes.io/projected/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-kube-api-access-64qlk\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.470796 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:06.970770098 +0000 UTC m=+152.666730650 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.571920 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.572919 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-registration-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.573147 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.073038608 +0000 UTC m=+152.768999160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573273 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-mountpoint-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573351 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-trusted-ca\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573381 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtjtn\" (UniqueName: \"kubernetes.io/projected/218bbc11-4dea-42a8-ab68-2ff2c2c84086-kube-api-access-rtjtn\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573451 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/956cc1d7-7e81-4e53-bae7-4adfe7139695-signing-cabundle\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573488 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hhcq2\" (UniqueName: \"kubernetes.io/projected/dead528a-47b6-44fc-97bf-17c308cc2033-kube-api-access-hhcq2\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573517 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7c13160-2a5b-463a-9216-f26d9143d19d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573549 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573579 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/08ce1e4a-c17a-4449-8cde-5febc97c501a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-shfrr\" (UID: \"08ce1e4a-c17a-4449-8cde-5febc97c501a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573648 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zlgs4\" (UniqueName: \"kubernetes.io/projected/2d42f253-2791-4b93-b572-ed832cbcdc1c-kube-api-access-zlgs4\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573675 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00806805-c6d3-4b27-aff6-3a6f6d03a094-trusted-ca\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573716 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2d42f253-2791-4b93-b572-ed832cbcdc1c-profile-collector-cert\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573755 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-socket-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.573822 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576156 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-csi-data-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576231 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/47afd50c-ef19-48a3-a7ad-98d6141719ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576288 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqtwc\" (UniqueName: \"kubernetes.io/projected/4cf5e708-ff2f-4b3c-9653-1aace2d4bfff-kube-api-access-vqtwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-gkbcz\" (UID: \"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576384 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00806805-c6d3-4b27-aff6-3a6f6d03a094-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576520 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576555 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-stats-auth\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576666 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-bound-sa-token\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576725 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/176f94a9-883c-4412-8ae3-6720c98cd4c1-certs\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576758 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2639202-023f-40ba-ad6e-291bc051dabe-config-volume\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576790 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bv866\" (UniqueName: \"kubernetes.io/projected/176f94a9-883c-4412-8ae3-6720c98cd4c1-kube-api-access-bv866\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576845 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rl7fm\" (UniqueName: \"kubernetes.io/projected/00806805-c6d3-4b27-aff6-3a6f6d03a094-kube-api-access-rl7fm\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576900 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-config\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576937 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tjdfx\" (UniqueName: \"kubernetes.io/projected/83013b2e-9ab3-4933-978b-9659e81810a9-kube-api-access-tjdfx\") pod \"ingress-canary-xk6gr\" (UID: \"83013b2e-9ab3-4933-978b-9659e81810a9\") " pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.576966 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb106cf7-e936-4502-a3bf-d154be0d6881-apiservice-cert\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577002 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-metrics-certs\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577031 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2639202-023f-40ba-ad6e-291bc051dabe-secret-volume\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577093 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ppp9h\" (UniqueName: \"kubernetes.io/projected/e207f660-e7ee-4ad4-abe2-c2037c1af333-kube-api-access-ppp9h\") pod \"package-server-manager-789f6589d5-bztls\" (UID: \"e207f660-e7ee-4ad4-abe2-c2037c1af333\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577141 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zmbx8\" (UniqueName: \"kubernetes.io/projected/08ce1e4a-c17a-4449-8cde-5febc97c501a-kube-api-access-zmbx8\") pod \"multus-admission-controller-857f4d67dd-shfrr\" (UID: \"08ce1e4a-c17a-4449-8cde-5febc97c501a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577138 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-trusted-ca\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577179 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-registry-tls\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577208 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-service-ca-bundle\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577335 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7wzmc\" (UniqueName: \"kubernetes.io/projected/68745ad4-99d4-4028-95fb-0d70e3fba519-kube-api-access-7wzmc\") pod \"migrator-59844c95c7-sb68p\" (UID: \"68745ad4-99d4-4028-95fb-0d70e3fba519\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577403 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-plugins-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577462 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-default-certificate\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577493 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64qlk\" (UniqueName: \"kubernetes.io/projected/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-kube-api-access-64qlk\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577522 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8394076a-e17a-4cea-a904-eb269c237f00-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577559 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c13160-2a5b-463a-9216-f26d9143d19d-config\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577594 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-serving-cert\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577656 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z997z\" (UniqueName: \"kubernetes.io/projected/956cc1d7-7e81-4e53-bae7-4adfe7139695-kube-api-access-z997z\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.578102 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-service-ca-bundle\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.581534 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.581731 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.081703957 +0000 UTC m=+152.777664509 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.582652 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.583528 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8394076a-e17a-4cea-a904-eb269c237f00-ca-trust-extracted\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.577699 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/949267c7-0029-430b-a356-7059518c503a-metrics-tls\") pod \"dns-operator-744455d44c-9wvt8\" (UID: \"949267c7-0029-430b-a356-7059518c503a\") " pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.584249 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb106cf7-e936-4502-a3bf-d154be0d6881-webhook-cert\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.584605 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cf5e708-ff2f-4b3c-9653-1aace2d4bfff-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gkbcz\" (UID: \"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.584711 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ckcxw\" (UniqueName: \"kubernetes.io/projected/bb106cf7-e936-4502-a3bf-d154be0d6881-kube-api-access-ckcxw\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585031 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7srcw\" (UniqueName: \"kubernetes.io/projected/949267c7-0029-430b-a356-7059518c503a-kube-api-access-7srcw\") pod \"dns-operator-744455d44c-9wvt8\" (UID: \"949267c7-0029-430b-a356-7059518c503a\") " pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585068 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/47afd50c-ef19-48a3-a7ad-98d6141719ec-srv-cert\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585095 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e207f660-e7ee-4ad4-abe2-c2037c1af333-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bztls\" (UID: \"e207f660-e7ee-4ad4-abe2-c2037c1af333\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585255 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8394076a-e17a-4cea-a904-eb269c237f00-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585387 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2d42f253-2791-4b93-b572-ed832cbcdc1c-srv-cert\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585454 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/176f94a9-883c-4412-8ae3-6720c98cd4c1-node-bootstrap-token\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585489 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/956cc1d7-7e81-4e53-bae7-4adfe7139695-signing-key\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585551 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vqw5\" (UniqueName: \"kubernetes.io/projected/d2639202-023f-40ba-ad6e-291bc051dabe-kube-api-access-6vqw5\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585586 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bb106cf7-e936-4502-a3bf-d154be0d6881-tmpfs\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.585650 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/218bbc11-4dea-42a8-ab68-2ff2c2c84086-config-volume\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.586761 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588051 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-55rk5\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-kube-api-access-55rk5\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588108 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-registry-certificates\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588159 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c13160-2a5b-463a-9216-f26d9143d19d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588231 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/00806805-c6d3-4b27-aff6-3a6f6d03a094-metrics-tls\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588285 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fn9r4\" (UniqueName: \"kubernetes.io/projected/47afd50c-ef19-48a3-a7ad-98d6141719ec-kube-api-access-fn9r4\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588318 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83013b2e-9ab3-4933-978b-9659e81810a9-cert\") pod \"ingress-canary-xk6gr\" (UID: \"83013b2e-9ab3-4933-978b-9659e81810a9\") " pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588344 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/218bbc11-4dea-42a8-ab68-2ff2c2c84086-metrics-tls\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588385 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vnbqd\" (UniqueName: \"kubernetes.io/projected/e089d4b5-58b6-41ed-9901-4d01c86024ab-kube-api-access-vnbqd\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.588420 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5hlpl\" (UniqueName: \"kubernetes.io/projected/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-kube-api-access-5hlpl\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.590619 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-metrics-certs\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.591175 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-registry-certificates\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.591241 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-stats-auth\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.592063 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/00806805-c6d3-4b27-aff6-3a6f6d03a094-trusted-ca\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.594515 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/949267c7-0029-430b-a356-7059518c503a-metrics-tls\") pod \"dns-operator-744455d44c-9wvt8\" (UID: \"949267c7-0029-430b-a356-7059518c503a\") " pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.623733 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8394076a-e17a-4cea-a904-eb269c237f00-installation-pull-secrets\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.624223 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-registry-tls\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.625848 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-default-certificate\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.627226 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/00806805-c6d3-4b27-aff6-3a6f6d03a094-metrics-tls\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.632674 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rl7fm\" (UniqueName: \"kubernetes.io/projected/00806805-c6d3-4b27-aff6-3a6f6d03a094-kube-api-access-rl7fm\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.644310 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-bound-sa-token\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.647783 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/00806805-c6d3-4b27-aff6-3a6f6d03a094-bound-sa-token\") pod \"ingress-operator-5b745b69d9-hs7zl\" (UID: \"00806805-c6d3-4b27-aff6-3a6f6d03a094\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.677299 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.678659 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hhcq2\" (UniqueName: \"kubernetes.io/projected/dead528a-47b6-44fc-97bf-17c308cc2033-kube-api-access-hhcq2\") pod \"marketplace-operator-79b997595-bwzxf\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.690181 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.690396 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691623 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c13160-2a5b-463a-9216-f26d9143d19d-config\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691678 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/c7c13160-2a5b-463a-9216-f26d9143d19d-config\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691725 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-serving-cert\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691751 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z997z\" (UniqueName: \"kubernetes.io/projected/956cc1d7-7e81-4e53-bae7-4adfe7139695-kube-api-access-z997z\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691772 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb106cf7-e936-4502-a3bf-d154be0d6881-webhook-cert\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691804 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cf5e708-ff2f-4b3c-9653-1aace2d4bfff-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gkbcz\" (UID: \"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691825 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ckcxw\" (UniqueName: \"kubernetes.io/projected/bb106cf7-e936-4502-a3bf-d154be0d6881-kube-api-access-ckcxw\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691850 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/47afd50c-ef19-48a3-a7ad-98d6141719ec-srv-cert\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691871 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e207f660-e7ee-4ad4-abe2-c2037c1af333-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bztls\" (UID: \"e207f660-e7ee-4ad4-abe2-c2037c1af333\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.691987 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2d42f253-2791-4b93-b572-ed832cbcdc1c-srv-cert\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692006 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/176f94a9-883c-4412-8ae3-6720c98cd4c1-node-bootstrap-token\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692024 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/956cc1d7-7e81-4e53-bae7-4adfe7139695-signing-key\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692045 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vqw5\" (UniqueName: \"kubernetes.io/projected/d2639202-023f-40ba-ad6e-291bc051dabe-kube-api-access-6vqw5\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692063 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bb106cf7-e936-4502-a3bf-d154be0d6881-tmpfs\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692086 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/218bbc11-4dea-42a8-ab68-2ff2c2c84086-config-volume\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692115 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c13160-2a5b-463a-9216-f26d9143d19d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692133 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fn9r4\" (UniqueName: \"kubernetes.io/projected/47afd50c-ef19-48a3-a7ad-98d6141719ec-kube-api-access-fn9r4\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692154 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83013b2e-9ab3-4933-978b-9659e81810a9-cert\") pod \"ingress-canary-xk6gr\" (UID: \"83013b2e-9ab3-4933-978b-9659e81810a9\") " pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692168 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/218bbc11-4dea-42a8-ab68-2ff2c2c84086-metrics-tls\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692186 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vnbqd\" (UniqueName: \"kubernetes.io/projected/e089d4b5-58b6-41ed-9901-4d01c86024ab-kube-api-access-vnbqd\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692205 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5hlpl\" (UniqueName: \"kubernetes.io/projected/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-kube-api-access-5hlpl\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692223 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-mountpoint-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692241 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-registration-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692261 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtjtn\" (UniqueName: \"kubernetes.io/projected/218bbc11-4dea-42a8-ab68-2ff2c2c84086-kube-api-access-rtjtn\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692281 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/956cc1d7-7e81-4e53-bae7-4adfe7139695-signing-cabundle\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692300 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7c13160-2a5b-463a-9216-f26d9143d19d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692322 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/08ce1e4a-c17a-4449-8cde-5febc97c501a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-shfrr\" (UID: \"08ce1e4a-c17a-4449-8cde-5febc97c501a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692342 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zlgs4\" (UniqueName: \"kubernetes.io/projected/2d42f253-2791-4b93-b572-ed832cbcdc1c-kube-api-access-zlgs4\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692360 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2d42f253-2791-4b93-b572-ed832cbcdc1c-profile-collector-cert\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692377 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-socket-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692397 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-csi-data-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692418 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/47afd50c-ef19-48a3-a7ad-98d6141719ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692848 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqtwc\" (UniqueName: \"kubernetes.io/projected/4cf5e708-ff2f-4b3c-9653-1aace2d4bfff-kube-api-access-vqtwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-gkbcz\" (UID: \"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692920 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/176f94a9-883c-4412-8ae3-6720c98cd4c1-certs\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692943 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2639202-023f-40ba-ad6e-291bc051dabe-config-volume\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692962 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bv866\" (UniqueName: \"kubernetes.io/projected/176f94a9-883c-4412-8ae3-6720c98cd4c1-kube-api-access-bv866\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.692985 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-config\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693006 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tjdfx\" (UniqueName: \"kubernetes.io/projected/83013b2e-9ab3-4933-978b-9659e81810a9-kube-api-access-tjdfx\") pod \"ingress-canary-xk6gr\" (UID: \"83013b2e-9ab3-4933-978b-9659e81810a9\") " pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693022 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb106cf7-e936-4502-a3bf-d154be0d6881-apiservice-cert\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693040 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2639202-023f-40ba-ad6e-291bc051dabe-secret-volume\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693049 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-registration-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693064 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zmbx8\" (UniqueName: \"kubernetes.io/projected/08ce1e4a-c17a-4449-8cde-5febc97c501a-kube-api-access-zmbx8\") pod \"multus-admission-controller-857f4d67dd-shfrr\" (UID: \"08ce1e4a-c17a-4449-8cde-5febc97c501a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693159 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ppp9h\" (UniqueName: \"kubernetes.io/projected/e207f660-e7ee-4ad4-abe2-c2037c1af333-kube-api-access-ppp9h\") pod \"package-server-manager-789f6589d5-bztls\" (UID: \"e207f660-e7ee-4ad4-abe2-c2037c1af333\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693238 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-plugins-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693418 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-plugins-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.693848 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/bb106cf7-e936-4502-a3bf-d154be0d6881-tmpfs\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.694460 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.194435507 +0000 UTC m=+152.890396059 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.694694 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/218bbc11-4dea-42a8-ab68-2ff2c2c84086-config-volume\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.695645 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/2d42f253-2791-4b93-b572-ed832cbcdc1c-srv-cert\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.696195 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/956cc1d7-7e81-4e53-bae7-4adfe7139695-signing-cabundle\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.696681 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/176f94a9-883c-4412-8ae3-6720c98cd4c1-node-bootstrap-token\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.697342 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-serving-cert\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.697362 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2639202-023f-40ba-ad6e-291bc051dabe-config-volume\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.697642 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-socket-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.697887 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/47afd50c-ef19-48a3-a7ad-98d6141719ec-srv-cert\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.698586 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7wzmc\" (UniqueName: \"kubernetes.io/projected/68745ad4-99d4-4028-95fb-0d70e3fba519-kube-api-access-7wzmc\") pod \"migrator-59844c95c7-sb68p\" (UID: \"68745ad4-99d4-4028-95fb-0d70e3fba519\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.698738 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/c7c13160-2a5b-463a-9216-f26d9143d19d-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.698938 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-config\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.699013 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-mountpoint-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.699405 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/e207f660-e7ee-4ad4-abe2-c2037c1af333-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-bztls\" (UID: \"e207f660-e7ee-4ad4-abe2-c2037c1af333\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.700053 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/08ce1e4a-c17a-4449-8cde-5febc97c501a-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-shfrr\" (UID: \"08ce1e4a-c17a-4449-8cde-5febc97c501a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.700138 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e089d4b5-58b6-41ed-9901-4d01c86024ab-csi-data-dir\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.701434 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/4cf5e708-ff2f-4b3c-9653-1aace2d4bfff-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-gkbcz\" (UID: \"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.701773 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/47afd50c-ef19-48a3-a7ad-98d6141719ec-profile-collector-cert\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.702071 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/bb106cf7-e936-4502-a3bf-d154be0d6881-apiservice-cert\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.702304 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/176f94a9-883c-4412-8ae3-6720c98cd4c1-certs\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.702980 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/956cc1d7-7e81-4e53-bae7-4adfe7139695-signing-key\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.703023 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/2d42f253-2791-4b93-b572-ed832cbcdc1c-profile-collector-cert\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.703185 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/bb106cf7-e936-4502-a3bf-d154be0d6881-webhook-cert\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.703807 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.704309 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/218bbc11-4dea-42a8-ab68-2ff2c2c84086-metrics-tls\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.705587 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/83013b2e-9ab3-4933-978b-9659e81810a9-cert\") pod \"ingress-canary-xk6gr\" (UID: \"83013b2e-9ab3-4933-978b-9659e81810a9\") " pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.711289 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.718949 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.721071 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2639202-023f-40ba-ad6e-291bc051dabe-secret-volume\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.722616 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64qlk\" (UniqueName: \"kubernetes.io/projected/ffef519e-8b2e-4a79-b15b-a6c3bb36a875-kube-api-access-64qlk\") pod \"router-default-5444994796-r245h\" (UID: \"ffef519e-8b2e-4a79-b15b-a6c3bb36a875\") " pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.726862 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.742539 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7srcw\" (UniqueName: \"kubernetes.io/projected/949267c7-0029-430b-a356-7059518c503a-kube-api-access-7srcw\") pod \"dns-operator-744455d44c-9wvt8\" (UID: \"949267c7-0029-430b-a356-7059518c503a\") " pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.746245 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.766836 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-55rk5\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-kube-api-access-55rk5\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.787413 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.807103 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.807527 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.307506607 +0000 UTC m=+153.003467159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.821984 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.825743 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z997z\" (UniqueName: \"kubernetes.io/projected/956cc1d7-7e81-4e53-bae7-4adfe7139695-kube-api-access-z997z\") pod \"service-ca-9c57cc56f-7txdb\" (UID: \"956cc1d7-7e81-4e53-bae7-4adfe7139695\") " pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.826515 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vqw5\" (UniqueName: \"kubernetes.io/projected/d2639202-023f-40ba-ad6e-291bc051dabe-kube-api-access-6vqw5\") pod \"collect-profiles-29402925-72d2x\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.830220 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.844848 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ckcxw\" (UniqueName: \"kubernetes.io/projected/bb106cf7-e936-4502-a3bf-d154be0d6881-kube-api-access-ckcxw\") pod \"packageserver-d55dfcdfc-6gfsk\" (UID: \"bb106cf7-e936-4502-a3bf-d154be0d6881\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.860966 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zlgs4\" (UniqueName: \"kubernetes.io/projected/2d42f253-2791-4b93-b572-ed832cbcdc1c-kube-api-access-zlgs4\") pod \"catalog-operator-68c6474976-62kvs\" (UID: \"2d42f253-2791-4b93-b572-ed832cbcdc1c\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.881846 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtjtn\" (UniqueName: \"kubernetes.io/projected/218bbc11-4dea-42a8-ab68-2ff2c2c84086-kube-api-access-rtjtn\") pod \"dns-default-vkf2v\" (UID: \"218bbc11-4dea-42a8-ab68-2ff2c2c84086\") " pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.907194 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq"] Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.909713 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.909946 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.409856389 +0000 UTC m=+153.105816951 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.910390 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:06 crc kubenswrapper[4956]: E1126 16:55:06.910921 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.410897649 +0000 UTC m=+153.106858201 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.920146 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/c7c13160-2a5b-463a-9216-f26d9143d19d-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-rjbgj\" (UID: \"c7c13160-2a5b-463a-9216-f26d9143d19d\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.923695 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zmbx8\" (UniqueName: \"kubernetes.io/projected/08ce1e4a-c17a-4449-8cde-5febc97c501a-kube-api-access-zmbx8\") pod \"multus-admission-controller-857f4d67dd-shfrr\" (UID: \"08ce1e4a-c17a-4449-8cde-5febc97c501a\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.938326 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqtwc\" (UniqueName: \"kubernetes.io/projected/4cf5e708-ff2f-4b3c-9653-1aace2d4bfff-kube-api-access-vqtwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-gkbcz\" (UID: \"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.968987 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vnbqd\" (UniqueName: \"kubernetes.io/projected/e089d4b5-58b6-41ed-9901-4d01c86024ab-kube-api-access-vnbqd\") pod \"csi-hostpathplugin-8htxc\" (UID: \"e089d4b5-58b6-41ed-9901-4d01c86024ab\") " pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:06 crc kubenswrapper[4956]: I1126 16:55:06.999511 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.012606 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.012765 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.512742137 +0000 UTC m=+153.208702689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.013036 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.013339 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.513328904 +0000 UTC m=+153.209289456 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.054322 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.060530 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.067746 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.072245 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tjdfx\" (UniqueName: \"kubernetes.io/projected/83013b2e-9ab3-4933-978b-9659e81810a9-kube-api-access-tjdfx\") pod \"ingress-canary-xk6gr\" (UID: \"83013b2e-9ab3-4933-978b-9659e81810a9\") " pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.072425 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bv866\" (UniqueName: \"kubernetes.io/projected/176f94a9-883c-4412-8ae3-6720c98cd4c1-kube-api-access-bv866\") pod \"machine-config-server-lh9p2\" (UID: \"176f94a9-883c-4412-8ae3-6720c98cd4c1\") " pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.074797 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.075258 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fn9r4\" (UniqueName: \"kubernetes.io/projected/47afd50c-ef19-48a3-a7ad-98d6141719ec-kube-api-access-fn9r4\") pod \"olm-operator-6b444d44fb-b9vfq\" (UID: \"47afd50c-ef19-48a3-a7ad-98d6141719ec\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.075574 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5hlpl\" (UniqueName: \"kubernetes.io/projected/850b303a-f8c1-4371-a7fb-a0e31cfe75d1-kube-api-access-5hlpl\") pod \"service-ca-operator-777779d784-24z97\" (UID: \"850b303a-f8c1-4371-a7fb-a0e31cfe75d1\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.078731 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ppp9h\" (UniqueName: \"kubernetes.io/projected/e207f660-e7ee-4ad4-abe2-c2037c1af333-kube-api-access-ppp9h\") pod \"package-server-manager-789f6589d5-bztls\" (UID: \"e207f660-e7ee-4ad4-abe2-c2037c1af333\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.082591 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.089650 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.096930 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.103628 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.114261 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.114379 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.614356298 +0000 UTC m=+153.310316850 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.114598 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.115011 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.615002136 +0000 UTC m=+153.310962688 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.127629 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.135319 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.142825 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-lh9p2" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.153195 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-xk6gr" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.157279 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.181498 4956 generic.go:334] "Generic (PLEG): container finished" podID="183ca9fa-6509-45fb-8f0a-41161814ecbb" containerID="adc2cec84e9d1fd5543b5003cf623e040506475a4f52c630ea372354eddd0b5b" exitCode=0 Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.181572 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" event={"ID":"183ca9fa-6509-45fb-8f0a-41161814ecbb","Type":"ContainerDied","Data":"adc2cec84e9d1fd5543b5003cf623e040506475a4f52c630ea372354eddd0b5b"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.185061 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" event={"ID":"ce09c66b-2af4-47e2-9f7d-bf540aca5b5e","Type":"ContainerStarted","Data":"a4677f85d122c44a3aa81468e16121ab5056ea3be7aaa556746ad0711c341436"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.197285 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" event={"ID":"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed","Type":"ContainerStarted","Data":"32b62d987f391af98ebc562d8376b8bd072753d96a3390295baf5d6e46347e89"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.197362 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" event={"ID":"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed","Type":"ContainerStarted","Data":"db91cb1fb22016f64d73facd702a2283ce091f26f725d9e48612c5c77d9d771d"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.197379 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" event={"ID":"6aa20dbb-4a6d-4a96-bf62-7c4e337285ed","Type":"ContainerStarted","Data":"79831c52215f60e9c983ef6b753dc60d153f06f8f011407c3dc96d933909446c"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.215850 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.216649 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.716594407 +0000 UTC m=+153.412554959 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.226297 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.234400 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" event={"ID":"706f4561-843f-4b61-bede-4d4c02d8dad6","Type":"ContainerStarted","Data":"6ff8c66cc50e196b50b2d111478c91ebab19d19ccb85c30952bfa40277781faa"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.270745 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" event={"ID":"be8528e4-668a-48fb-b8b9-556f9d57118a","Type":"ContainerStarted","Data":"df140d89ffb8be4f1efedd94cd6287603c00ac35f0a2f5840504d443e425c8d4"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.270821 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" event={"ID":"be8528e4-668a-48fb-b8b9-556f9d57118a","Type":"ContainerStarted","Data":"d0025771789b862adcf297666a5c1adcf5013c41aa53356597aae2f3f356ae71"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.270844 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.283357 4956 patch_prober.go:28] interesting pod/console-operator-58897d9998-nd5x4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.284299 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" podUID="be8528e4-668a-48fb-b8b9-556f9d57118a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.285238 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r245h" event={"ID":"ffef519e-8b2e-4a79-b15b-a6c3bb36a875","Type":"ContainerStarted","Data":"be7708c830a54ddeb6f390b0e490d21645cceeaf380bf9fb54bd5b183563dcd2"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.286848 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" event={"ID":"b1fae2fb-ef9b-4085-b673-7514924aa516","Type":"ContainerStarted","Data":"d13e148439d2d48bb26331743aa8b96e37de9620f4b684e9858c70d3ce3d5fc0"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.290553 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" event={"ID":"1761b713-5b59-4758-99cc-11a2e8eff4ae","Type":"ContainerStarted","Data":"9d445cacee49844346f5f9ff60fbd29b527964f1eadebb878499fec233c51d95"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.292210 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-cwlp6" event={"ID":"996f361a-f4aa-4d79-b12c-3e357f126f43","Type":"ContainerStarted","Data":"b8ca71d7517a7c12ee9d54469176f1b663972613655c69b74fa39b5b207ef807"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.294031 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.308943 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" event={"ID":"724bfa7f-010c-4298-a11a-50ea0f5fe198","Type":"ContainerStarted","Data":"8f87abf1bc1d3186dc6a7ddfc4d5069a4359687e3eab741db40e73e037eff03c"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.308993 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" event={"ID":"724bfa7f-010c-4298-a11a-50ea0f5fe198","Type":"ContainerStarted","Data":"69d67044deda3a6f8ff63fe22e43e5929a1792804a7fa91695285290cbfc7d31"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.309008 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" event={"ID":"724bfa7f-010c-4298-a11a-50ea0f5fe198","Type":"ContainerStarted","Data":"2696acae08f95ca76d2b69baf7c183f681a618bce862f31b60d1bd2f0a0bfca2"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.309771 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" event={"ID":"ce58bfc7-f777-4db0-8690-ce9a4137a92c","Type":"ContainerStarted","Data":"9a6289092a4c4c68a318564159b49d7ac848a541dc4d00731fbf0226e85467c8"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.311282 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" event={"ID":"f3f11c7e-1585-49d4-8675-2b4234fd005e","Type":"ContainerStarted","Data":"46d9525e03f77e22b58fc24d71550a4db7f1cf38215c283b9e4990d988850b53"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.311344 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" event={"ID":"f3f11c7e-1585-49d4-8675-2b4234fd005e","Type":"ContainerStarted","Data":"7dcc276db7070f1e67ad54376764934aa41c41727e55e78537e864d96ba4b412"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.314629 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" event={"ID":"5552a148-c82e-4afb-9a27-9ce40714bdb0","Type":"ContainerStarted","Data":"7d3d6a4e7e389b2af8fdeea87b1322ddd0313dcdfc2f9f43ada86d801a671cd8"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.314659 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" event={"ID":"5552a148-c82e-4afb-9a27-9ce40714bdb0","Type":"ContainerStarted","Data":"59f9a0880b9fdd99828564611d7f38e3ce6e8f06e2baf458618a0a1b55089705"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.315478 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.315594 4956 patch_prober.go:28] interesting pod/downloads-7954f5f757-cwlp6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.315635 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-cwlp6" podUID="996f361a-f4aa-4d79-b12c-3e357f126f43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.317228 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.328319 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-ssdwk" event={"ID":"166d8e6f-9af9-4212-9593-0b47d27918d1","Type":"ContainerStarted","Data":"6e98f8580622a8c349a3469ef3f9f2e532e6a29d2251e088abfd1d9cd20e348f"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.329563 4956 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8tmqv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.28:6443/healthz\": dial tcp 10.217.0.28:6443: connect: connection refused" start-of-body= Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.329656 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.28:6443/healthz\": dial tcp 10.217.0.28:6443: connect: connection refused" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.335892 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.835842304 +0000 UTC m=+153.531802856 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.338112 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" event={"ID":"5e23939f-1541-494e-af81-1cf62ff004b0","Type":"ContainerStarted","Data":"1a348aee14c1f3a1a280bd7f7e1949b2d282f44dc95755b8db795cd0b56b6e68"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.354360 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwzxf"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.357002 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" event={"ID":"8dd2caa0-c053-4d2b-9b22-dfbf133065ac","Type":"ContainerStarted","Data":"e072e1ea2bc722bd684024edbbe08e107625d8c2c177b8ada0d1a18a04e7a5cb"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.357087 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" event={"ID":"8dd2caa0-c053-4d2b-9b22-dfbf133065ac","Type":"ContainerStarted","Data":"762f0aceead5162441f3596db66fef7ab669d7dd283bbe0928dc14587b38a5fd"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.361680 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.362790 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" event={"ID":"b9b65044-9326-4a72-a933-e84012a29211","Type":"ContainerStarted","Data":"7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.363174 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.365057 4956 generic.go:334] "Generic (PLEG): container finished" podID="bea9e494-4444-40f7-8b17-f5ee6293d553" containerID="19eac008eaadf0a50b4215e11d27308d2ae46d8acaebbb55e3acbc3cbe3bf5dc" exitCode=0 Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.365221 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" event={"ID":"bea9e494-4444-40f7-8b17-f5ee6293d553","Type":"ContainerDied","Data":"19eac008eaadf0a50b4215e11d27308d2ae46d8acaebbb55e3acbc3cbe3bf5dc"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.365302 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" event={"ID":"bea9e494-4444-40f7-8b17-f5ee6293d553","Type":"ContainerStarted","Data":"9f017a03c265ab57929c2dcf9d8bfc42eacafa9f6b4f7e30b8e51568b91b60ac"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.366518 4956 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-v5xlx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.368064 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" podUID="b9b65044-9326-4a72-a933-e84012a29211" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.371214 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.371936 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" event={"ID":"fec6a02f-63fc-473e-8a99-fcbe8ce154a6","Type":"ContainerStarted","Data":"fadea983c5d37369dd564e90f2b2be27ef1c65c073dd21063582e0837afe6854"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.372021 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" event={"ID":"fec6a02f-63fc-473e-8a99-fcbe8ce154a6","Type":"ContainerStarted","Data":"cc69824bd839013b1bab8e2b83baffb44a272ae240af65403829f5b4fbfd056f"} Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.377486 4956 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-kr7n2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.377548 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.15:8443/healthz\": dial tcp 10.217.0.15:8443: connect: connection refused" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.378710 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.434077 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.434302 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.934257013 +0000 UTC m=+153.630217575 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.435000 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.437140 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:07.937120806 +0000 UTC m=+153.633081548 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.477556 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.481767 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt"] Nov 26 16:55:07 crc kubenswrapper[4956]: W1126 16:55:07.494124 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod00806805_c6d3_4b27_aff6_3a6f6d03a094.slice/crio-9e1c61a71221840ce12ed11a95e0bf77966475639c0ee0206662a80e3db34434 WatchSource:0}: Error finding container 9e1c61a71221840ce12ed11a95e0bf77966475639c0ee0206662a80e3db34434: Status 404 returned error can't find the container with id 9e1c61a71221840ce12ed11a95e0bf77966475639c0ee0206662a80e3db34434 Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.497591 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs"] Nov 26 16:55:07 crc kubenswrapper[4956]: W1126 16:55:07.506354 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podae2ec484_4014_4235_bbe5_aa47e9f0f083.slice/crio-b92349c85a227097938960bf456d22d25e9ca3e68629cd7b23ac173283e3a92f WatchSource:0}: Error finding container b92349c85a227097938960bf456d22d25e9ca3e68629cd7b23ac173283e3a92f: Status 404 returned error can't find the container with id b92349c85a227097938960bf456d22d25e9ca3e68629cd7b23ac173283e3a92f Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.521339 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9wvt8"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.535734 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.538266 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.038235812 +0000 UTC m=+153.734196524 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: W1126 16:55:07.569237 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9a53de27_4165_434e_8f23_6180d37c530c.slice/crio-803805c75a4381513f15c2af22b88e6f5224aae3411dc7be199dcfd966d4a5b9 WatchSource:0}: Error finding container 803805c75a4381513f15c2af22b88e6f5224aae3411dc7be199dcfd966d4a5b9: Status 404 returned error can't find the container with id 803805c75a4381513f15c2af22b88e6f5224aae3411dc7be199dcfd966d4a5b9 Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.575829 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj"] Nov 26 16:55:07 crc kubenswrapper[4956]: W1126 16:55:07.631655 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod949267c7_0029_430b_a356_7059518c503a.slice/crio-512bd7444412372af8cc841afa1a134b9cc1a9e3ccc2428305e171b5ebe681d2 WatchSource:0}: Error finding container 512bd7444412372af8cc841afa1a134b9cc1a9e3ccc2428305e171b5ebe681d2: Status 404 returned error can't find the container with id 512bd7444412372af8cc841afa1a134b9cc1a9e3ccc2428305e171b5ebe681d2 Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.639121 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.639567 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.139549395 +0000 UTC m=+153.835509947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.692804 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-shfrr"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.741786 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.742069 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.241995439 +0000 UTC m=+153.937955991 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.742309 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.742908 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.242891375 +0000 UTC m=+153.938851927 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.747668 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-z5f2l" podStartSLOduration=127.747640942 podStartE2EDuration="2m7.747640942s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:07.723356104 +0000 UTC m=+153.419316666" watchObservedRunningTime="2025-11-26 16:55:07.747640942 +0000 UTC m=+153.443601484" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.849755 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz"] Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.849982 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.850063 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.350043125 +0000 UTC m=+154.046003677 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.851049 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.851508 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.351493507 +0000 UTC m=+154.047454059 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.914738 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-7vqnn" podStartSLOduration=127.914709954 podStartE2EDuration="2m7.914709954s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:07.909753702 +0000 UTC m=+153.605714254" watchObservedRunningTime="2025-11-26 16:55:07.914709954 +0000 UTC m=+153.610670496" Nov 26 16:55:07 crc kubenswrapper[4956]: I1126 16:55:07.952727 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:07 crc kubenswrapper[4956]: E1126 16:55:07.953583 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.453559721 +0000 UTC m=+154.149520273 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.015745 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" podStartSLOduration=128.015707807 podStartE2EDuration="2m8.015707807s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.015532062 +0000 UTC m=+153.711492634" watchObservedRunningTime="2025-11-26 16:55:08.015707807 +0000 UTC m=+153.711668359" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.060669 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.061216 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.561196395 +0000 UTC m=+154.257156947 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.081165 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-dl28z" podStartSLOduration=128.081146808 podStartE2EDuration="2m8.081146808s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.080011146 +0000 UTC m=+153.775971698" watchObservedRunningTime="2025-11-26 16:55:08.081146808 +0000 UTC m=+153.777107360" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.135116 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-8htxc"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.140849 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.147179 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-874qk" podStartSLOduration=128.147163936 podStartE2EDuration="2m8.147163936s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.124441233 +0000 UTC m=+153.820401795" watchObservedRunningTime="2025-11-26 16:55:08.147163936 +0000 UTC m=+153.843124488" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.148220 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.162934 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.163391 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.663371512 +0000 UTC m=+154.359332064 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.165615 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" podStartSLOduration=128.165601546 podStartE2EDuration="2m8.165601546s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.1636586 +0000 UTC m=+153.859619152" watchObservedRunningTime="2025-11-26 16:55:08.165601546 +0000 UTC m=+153.861562098" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.174304 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.188998 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-vkf2v"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.208627 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-24z97"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.209399 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-cwlp6" podStartSLOduration=128.209379034 podStartE2EDuration="2m8.209379034s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.205722009 +0000 UTC m=+153.901682571" watchObservedRunningTime="2025-11-26 16:55:08.209379034 +0000 UTC m=+153.905339586" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.246528 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-7txdb"] Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.267992 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.268974 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.768927666 +0000 UTC m=+154.464888218 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.299285 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-xk6gr"] Nov 26 16:55:08 crc kubenswrapper[4956]: W1126 16:55:08.323333 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode207f660_e7ee_4ad4_abe2_c2037c1af333.slice/crio-28a581906acf493ba0ea60fd06f31f6ce9209e95eefeec815de8d5eb4b349b1a WatchSource:0}: Error finding container 28a581906acf493ba0ea60fd06f31f6ce9209e95eefeec815de8d5eb4b349b1a: Status 404 returned error can't find the container with id 28a581906acf493ba0ea60fd06f31f6ce9209e95eefeec815de8d5eb4b349b1a Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.369383 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.369859 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.869828277 +0000 UTC m=+154.565788819 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.446980 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" event={"ID":"2d42f253-2791-4b93-b572-ed832cbcdc1c","Type":"ContainerStarted","Data":"3f9de3216eceef16defb988e777de4bd00abdae88a2d0bb5ee5471b86f6d4f5e"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.447023 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" event={"ID":"2d42f253-2791-4b93-b572-ed832cbcdc1c","Type":"ContainerStarted","Data":"dd814f10f5b74c87e81a8842297fcb1a2b81286e7276d692dbf19a5df5a2d06a"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.449838 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" event={"ID":"ce58bfc7-f777-4db0-8690-ce9a4137a92c","Type":"ContainerStarted","Data":"f8410c838d13295e8cac7e6d59a3e44a469bcb05d9a0256c7084e380d6e6425d"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.468511 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" event={"ID":"68745ad4-99d4-4028-95fb-0d70e3fba519","Type":"ContainerStarted","Data":"3c4ed26ca744564f723f25fba0cdf81fb1dd926a2c7fee2aa5cdaf90c3dc52ac"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.473662 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.474277 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:08.974260888 +0000 UTC m=+154.670221440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.485071 4956 generic.go:334] "Generic (PLEG): container finished" podID="5e23939f-1541-494e-af81-1cf62ff004b0" containerID="691aebd9d35a0ffa55811f895d729232909526eb66d058b4fdded67f17327997" exitCode=0 Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.485139 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" event={"ID":"5e23939f-1541-494e-af81-1cf62ff004b0","Type":"ContainerDied","Data":"691aebd9d35a0ffa55811f895d729232909526eb66d058b4fdded67f17327997"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.492847 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" event={"ID":"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff","Type":"ContainerStarted","Data":"f7e97eb2f3425561d81320de341d58a80baa89b71366e5531ecad5dc8cf1c364"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.493141 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-j2x66" podStartSLOduration=128.493130361 podStartE2EDuration="2m8.493130361s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.464334873 +0000 UTC m=+154.160295425" watchObservedRunningTime="2025-11-26 16:55:08.493130361 +0000 UTC m=+154.189090903" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.495360 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" event={"ID":"bb106cf7-e936-4502-a3bf-d154be0d6881","Type":"ContainerStarted","Data":"4a4110645afa9da751b7e0b22de19ea267259fa59ae290d7ea8acfa849a63149"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.499453 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" event={"ID":"08ce1e4a-c17a-4449-8cde-5febc97c501a","Type":"ContainerStarted","Data":"94010de4d4cbb24e283051c5bbb5272eca7f597e66ed07aef4db225a869fa921"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.500380 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" event={"ID":"ae2ec484-4014-4235-bbe5-aa47e9f0f083","Type":"ContainerStarted","Data":"b92349c85a227097938960bf456d22d25e9ca3e68629cd7b23ac173283e3a92f"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.511514 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" event={"ID":"00806805-c6d3-4b27-aff6-3a6f6d03a094","Type":"ContainerStarted","Data":"b1120900867f8c8132952ade9ff7fb9196248a8f8f4fe261c7fe65bdccfb35db"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.511552 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" event={"ID":"00806805-c6d3-4b27-aff6-3a6f6d03a094","Type":"ContainerStarted","Data":"9e1c61a71221840ce12ed11a95e0bf77966475639c0ee0206662a80e3db34434"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.516267 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lh9p2" event={"ID":"176f94a9-883c-4412-8ae3-6720c98cd4c1","Type":"ContainerStarted","Data":"52217ea17272b021d7f93954117243bbddffe1f7717b17e5225aff395c46866a"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.518894 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" event={"ID":"47afd50c-ef19-48a3-a7ad-98d6141719ec","Type":"ContainerStarted","Data":"9f5cd971e26c296d4cf36e107e3c47217a0e2bebe1692f6aecc2404b2780de29"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.520823 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" event={"ID":"e089d4b5-58b6-41ed-9901-4d01c86024ab","Type":"ContainerStarted","Data":"db78ac23c8052b92754d0c8b58deffe1afe7203b4e49640f03a0119c9ad92723"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.523434 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" event={"ID":"9a53de27-4165-434e-8f23-6180d37c530c","Type":"ContainerStarted","Data":"75547ce837d1d746669fc8bf9060b6eeb5c333b71fc1adb2ecd709ec5111b714"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.523465 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" event={"ID":"9a53de27-4165-434e-8f23-6180d37c530c","Type":"ContainerStarted","Data":"803805c75a4381513f15c2af22b88e6f5224aae3411dc7be199dcfd966d4a5b9"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.525174 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" event={"ID":"706f4561-843f-4b61-bede-4d4c02d8dad6","Type":"ContainerStarted","Data":"77430c0133c7f6779f3c3e42cb32f89da098148d848c0692d8340eb79118e284"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.528010 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" event={"ID":"c7c13160-2a5b-463a-9216-f26d9143d19d","Type":"ContainerStarted","Data":"c74e33b656b31aee1828992b78fe958760129e1d12bcc8575613bd12a6efa6cf"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.554402 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" event={"ID":"dead528a-47b6-44fc-97bf-17c308cc2033","Type":"ContainerStarted","Data":"16eb9510dfe9505048b0ed0c5b9f2fedb403dc8428a54678ac215f66f4342223"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.564879 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" event={"ID":"bea9e494-4444-40f7-8b17-f5ee6293d553","Type":"ContainerStarted","Data":"eb9f5082fc1aebd40f0df43194892e5fdd570fd36237e1e3e88f65d8d8ee6fbc"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.565476 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.575589 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-r245h" event={"ID":"ffef519e-8b2e-4a79-b15b-a6c3bb36a875","Type":"ContainerStarted","Data":"6e7afb7c9f1e6374037d0b5223b834d2c17d7572bb45c712a1ffaf1c6a57a2fd"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.577518 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" event={"ID":"229601a4-1c80-4ed2-9330-4297573d9af6","Type":"ContainerStarted","Data":"ece2d8dd7bf3e57e490cb6e1fa435361398af4a202cb1918b8ebfe3e236d89f5"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.581509 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.582537 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.08250289 +0000 UTC m=+154.778463442 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.583855 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.586436 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.086416912 +0000 UTC m=+154.782377464 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.622840 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" event={"ID":"949267c7-0029-430b-a356-7059518c503a","Type":"ContainerStarted","Data":"512bd7444412372af8cc841afa1a134b9cc1a9e3ccc2428305e171b5ebe681d2"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.635540 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" event={"ID":"d2639202-023f-40ba-ad6e-291bc051dabe","Type":"ContainerStarted","Data":"146167b02e819d75cf94b8c1b3dacc13296597bdce70399513d1a8c375934ab7"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.649516 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" event={"ID":"e207f660-e7ee-4ad4-abe2-c2037c1af333","Type":"ContainerStarted","Data":"28a581906acf493ba0ea60fd06f31f6ce9209e95eefeec815de8d5eb4b349b1a"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.671503 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" podStartSLOduration=128.671481218 podStartE2EDuration="2m8.671481218s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.626224707 +0000 UTC m=+154.322185259" watchObservedRunningTime="2025-11-26 16:55:08.671481218 +0000 UTC m=+154.367441760" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.685373 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" event={"ID":"1761b713-5b59-4758-99cc-11a2e8eff4ae","Type":"ContainerStarted","Data":"4a34833d75f910d0b627d02984083deecc9fd38e008a4ce7697f35494f4c84fe"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.698650 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.700144 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.200116441 +0000 UTC m=+154.896076993 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.709538 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-ssdwk" podStartSLOduration=128.709515191 podStartE2EDuration="2m8.709515191s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.705443444 +0000 UTC m=+154.401403996" watchObservedRunningTime="2025-11-26 16:55:08.709515191 +0000 UTC m=+154.405475743" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.730210 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.730826 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" start-of-body= Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.730903 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="Get \"http://localhost:1936/healthz/ready\": dial tcp [::1]:1936: connect: connection refused" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.742985 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" podStartSLOduration=128.742965032 podStartE2EDuration="2m8.742965032s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.741057628 +0000 UTC m=+154.437018180" watchObservedRunningTime="2025-11-26 16:55:08.742965032 +0000 UTC m=+154.438925584" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.758164 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" event={"ID":"850b303a-f8c1-4371-a7fb-a0e31cfe75d1","Type":"ContainerStarted","Data":"5285237d4b8e2b0c473ab250c5dfbca3bd2917bfbe33198a1a5593cc17f58fcf"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.778655 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vkf2v" event={"ID":"218bbc11-4dea-42a8-ab68-2ff2c2c84086","Type":"ContainerStarted","Data":"4078b41439fce609a463d26326928f5479f3f6f2428fd182e33d0c0f97c3d67f"} Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779439 4956 patch_prober.go:28] interesting pod/console-operator-58897d9998-nd5x4 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" start-of-body= Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779466 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" podUID="be8528e4-668a-48fb-b8b9-556f9d57118a" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.19:8443/readyz\": dial tcp 10.217.0.19:8443: connect: connection refused" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779607 4956 patch_prober.go:28] interesting pod/downloads-7954f5f757-cwlp6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779648 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-cwlp6" podUID="996f361a-f4aa-4d79-b12c-3e357f126f43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779746 4956 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-8tmqv container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.28:6443/healthz\": dial tcp 10.217.0.28:6443: connect: connection refused" start-of-body= Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779765 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.28:6443/healthz\": dial tcp 10.217.0.28:6443: connect: connection refused" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779902 4956 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-v5xlx container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.779923 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" podUID="b9b65044-9326-4a72-a933-e84012a29211" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.807489 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.807986 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.307967301 +0000 UTC m=+155.003927853 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.860473 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-v8mzp" podStartSLOduration=128.86045141 podStartE2EDuration="2m8.86045141s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:08.8583681 +0000 UTC m=+154.554328662" watchObservedRunningTime="2025-11-26 16:55:08.86045141 +0000 UTC m=+154.556411952" Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.908949 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.909235 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.409201751 +0000 UTC m=+155.105162303 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:08 crc kubenswrapper[4956]: I1126 16:55:08.909439 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:08 crc kubenswrapper[4956]: E1126 16:55:08.912949 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.412939428 +0000 UTC m=+155.108899980 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.018519 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.019371 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.519349187 +0000 UTC m=+155.215309739 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.065451 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-bpv5w" podStartSLOduration=129.065421781 podStartE2EDuration="2m9.065421781s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.064721741 +0000 UTC m=+154.760682293" watchObservedRunningTime="2025-11-26 16:55:09.065421781 +0000 UTC m=+154.761382333" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.162554 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.163099 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.663076909 +0000 UTC m=+155.359037461 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.261221 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-r245h" podStartSLOduration=129.261201689 podStartE2EDuration="2m9.261201689s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.25912782 +0000 UTC m=+154.955088372" watchObservedRunningTime="2025-11-26 16:55:09.261201689 +0000 UTC m=+154.957162241" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.264252 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.264772 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.764663489 +0000 UTC m=+155.460624051 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.264854 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.265224 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.765210604 +0000 UTC m=+155.461171156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.365198 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.365371 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.865330532 +0000 UTC m=+155.561291084 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.365666 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.366026 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.866018542 +0000 UTC m=+155.561979094 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.377584 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-z4gz4" podStartSLOduration=129.377565514 podStartE2EDuration="2m9.377565514s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.304420371 +0000 UTC m=+155.000380923" watchObservedRunningTime="2025-11-26 16:55:09.377565514 +0000 UTC m=+155.073526066" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.415369 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" podStartSLOduration=129.41534832 podStartE2EDuration="2m9.41534832s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.377971056 +0000 UTC m=+155.073931668" watchObservedRunningTime="2025-11-26 16:55:09.41534832 +0000 UTC m=+155.111308872" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.472410 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.473122 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:09.973103069 +0000 UTC m=+155.669063621 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.477218 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-jwqpq" podStartSLOduration=129.477201747 podStartE2EDuration="2m9.477201747s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.472226514 +0000 UTC m=+155.168187066" watchObservedRunningTime="2025-11-26 16:55:09.477201747 +0000 UTC m=+155.173162299" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.564534 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.564594 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.587434 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.587752 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.087740415 +0000 UTC m=+155.783700967 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.689271 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.690365 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.190342194 +0000 UTC m=+155.886302746 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.734800 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:09 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:09 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:09 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.734855 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.791920 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.793053 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.293040316 +0000 UTC m=+155.989000868 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.820514 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" event={"ID":"d2639202-023f-40ba-ad6e-291bc051dabe","Type":"ContainerStarted","Data":"0b47c13d2907c7dc393d82565da781ef0ea33cfa6c444f902cf2534ef72db73f"} Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.858593 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" event={"ID":"ae2ec484-4014-4235-bbe5-aa47e9f0f083","Type":"ContainerStarted","Data":"2f7697f445c4d46b9ed1a0d948535b14e6033e03c54dc3e87924cee1a157a4fa"} Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.863928 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" event={"ID":"c7c13160-2a5b-463a-9216-f26d9143d19d","Type":"ContainerStarted","Data":"6a50cdb9674b524246295e62d194f061157be1ab41d772cb7e0493673b0d7354"} Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.905828 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" podStartSLOduration=129.905809767 podStartE2EDuration="2m9.905809767s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.905479858 +0000 UTC m=+155.601440410" watchObservedRunningTime="2025-11-26 16:55:09.905809767 +0000 UTC m=+155.601770319" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.905885 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:09 crc kubenswrapper[4956]: E1126 16:55:09.909447 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.409431032 +0000 UTC m=+156.105391584 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.935046 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" event={"ID":"956cc1d7-7e81-4e53-bae7-4adfe7139695","Type":"ContainerStarted","Data":"35091bc7c520f9fe236e447f875ff94800ebb9f7e1757f426b3cbd18dda68b17"} Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.962866 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-dbctb" podStartSLOduration=129.962840697 podStartE2EDuration="2m9.962840697s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.948753792 +0000 UTC m=+155.644714344" watchObservedRunningTime="2025-11-26 16:55:09.962840697 +0000 UTC m=+155.658801249" Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.972131 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-lh9p2" event={"ID":"176f94a9-883c-4412-8ae3-6720c98cd4c1","Type":"ContainerStarted","Data":"2ce138a8752056eaed94e6cc9adf9211eda297ee2ec93e5aa34925b7d6961b7f"} Nov 26 16:55:09 crc kubenswrapper[4956]: I1126 16:55:09.994418 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-rjbgj" podStartSLOduration=129.994396784 podStartE2EDuration="2m9.994396784s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:09.992928572 +0000 UTC m=+155.688889134" watchObservedRunningTime="2025-11-26 16:55:09.994396784 +0000 UTC m=+155.690357336" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.004464 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" event={"ID":"229601a4-1c80-4ed2-9330-4297573d9af6","Type":"ContainerStarted","Data":"c614032453548838bcf75d2d1aa4397fe3b3da07c8b2847119b5046d3b9d198b"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.008335 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.010250 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.510234569 +0000 UTC m=+156.206195121 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.031621 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-lh9p2" podStartSLOduration=7.031597843 podStartE2EDuration="7.031597843s" podCreationTimestamp="2025-11-26 16:55:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.027156956 +0000 UTC m=+155.723117518" watchObservedRunningTime="2025-11-26 16:55:10.031597843 +0000 UTC m=+155.727558395" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.055277 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" event={"ID":"9a53de27-4165-434e-8f23-6180d37c530c","Type":"ContainerStarted","Data":"8fb8b209d5ec19c612ab35ce79fc078418b4e8b6ecd54ec869fcc9b8252b9a65"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.076729 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-tq7kg" podStartSLOduration=130.076688019 podStartE2EDuration="2m10.076688019s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.073544319 +0000 UTC m=+155.769504891" watchObservedRunningTime="2025-11-26 16:55:10.076688019 +0000 UTC m=+155.772648571" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.109456 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.110196 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.610177682 +0000 UTC m=+156.306138234 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.113458 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" event={"ID":"706f4561-843f-4b61-bede-4d4c02d8dad6","Type":"ContainerStarted","Data":"bfe1fe4c7f7992b9a59f92f4ba80a5de23ebcdce799826deca73a3e608dc86d2"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.194153 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-tn8vt" podStartSLOduration=130.194127375 podStartE2EDuration="2m10.194127375s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.142258954 +0000 UTC m=+155.838219506" watchObservedRunningTime="2025-11-26 16:55:10.194127375 +0000 UTC m=+155.890087927" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.213251 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" event={"ID":"949267c7-0029-430b-a356-7059518c503a","Type":"ContainerStarted","Data":"ac19305a4997db3fd432bb9a90f000343fabed9503a8e3e2331258666bfd2f54"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.215396 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.217302 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.717285341 +0000 UTC m=+156.413245893 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.225786 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-6mkc4" podStartSLOduration=130.225767315 podStartE2EDuration="2m10.225767315s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.194092284 +0000 UTC m=+155.890052846" watchObservedRunningTime="2025-11-26 16:55:10.225767315 +0000 UTC m=+155.921727867" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.260791 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" event={"ID":"dead528a-47b6-44fc-97bf-17c308cc2033","Type":"ContainerStarted","Data":"f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.261280 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.295786 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" podStartSLOduration=130.295759127 podStartE2EDuration="2m10.295759127s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.294409168 +0000 UTC m=+155.990369730" watchObservedRunningTime="2025-11-26 16:55:10.295759127 +0000 UTC m=+155.991719669" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.314296 4956 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bwzxf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.314368 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.316420 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.316714 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.816664578 +0000 UTC m=+156.512625130 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.317023 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.317467 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.81745893 +0000 UTC m=+156.513419482 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.319103 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xk6gr" event={"ID":"83013b2e-9ab3-4933-978b-9659e81810a9","Type":"ContainerStarted","Data":"eed69165829d83dddec8cc2866957cc03c10ecdac2ec66a7508252f5dfaecfd5"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.319149 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-xk6gr" event={"ID":"83013b2e-9ab3-4933-978b-9659e81810a9","Type":"ContainerStarted","Data":"920fa74f7e2c484f73642a22a5229dfdb5ed557cf414d45e1025226b64f0035b"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.323676 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" event={"ID":"850b303a-f8c1-4371-a7fb-a0e31cfe75d1","Type":"ContainerStarted","Data":"a0986033c315b6a8fb97cc1daff127035fb6d39dbdc51baa77bebf5a691a85b6"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.349272 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" event={"ID":"183ca9fa-6509-45fb-8f0a-41161814ecbb","Type":"ContainerStarted","Data":"aaa3c33d9d3cd8311b129d4fd441a55c516d3ee8437c0e7df78cf183dcc181c0"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.354454 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" event={"ID":"68745ad4-99d4-4028-95fb-0d70e3fba519","Type":"ContainerStarted","Data":"11781a20c48bbc2dc000eb27c8cb372732e9983b124cf545fd4016f38f38daad"} Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.379935 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-xk6gr" podStartSLOduration=6.379902245 podStartE2EDuration="6.379902245s" podCreationTimestamp="2025-11-26 16:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.349602344 +0000 UTC m=+156.045562896" watchObservedRunningTime="2025-11-26 16:55:10.379902245 +0000 UTC m=+156.075862787" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.381153 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-24z97" podStartSLOduration=130.381146151 podStartE2EDuration="2m10.381146151s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.378217617 +0000 UTC m=+156.074178169" watchObservedRunningTime="2025-11-26 16:55:10.381146151 +0000 UTC m=+156.077106703" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.405978 4956 patch_prober.go:28] interesting pod/downloads-7954f5f757-cwlp6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.406028 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-cwlp6" podUID="996f361a-f4aa-4d79-b12c-3e357f126f43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.405984 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.419567 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.419662 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.919649628 +0000 UTC m=+156.615610180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.420479 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.423874 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.427154 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:10.927135603 +0000 UTC m=+156.623096155 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.445352 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.455352 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" podStartSLOduration=130.455334154 podStartE2EDuration="2m10.455334154s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.428359928 +0000 UTC m=+156.124320480" watchObservedRunningTime="2025-11-26 16:55:10.455334154 +0000 UTC m=+156.151294706" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.515125 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-62kvs" podStartSLOduration=130.515094772 podStartE2EDuration="2m10.515094772s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:10.457431294 +0000 UTC m=+156.153391846" watchObservedRunningTime="2025-11-26 16:55:10.515094772 +0000 UTC m=+156.211055324" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.531685 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.532791 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.032768559 +0000 UTC m=+156.728729111 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.635052 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.635366 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.135352388 +0000 UTC m=+156.831312940 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.735992 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.736308 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.236275929 +0000 UTC m=+156.932236481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.736528 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.736898 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.236852946 +0000 UTC m=+156.932813498 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.736475 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:10 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:10 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:10 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.737065 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.838453 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.838780 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.338762765 +0000 UTC m=+157.034723317 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:10 crc kubenswrapper[4956]: I1126 16:55:10.940382 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:10 crc kubenswrapper[4956]: E1126 16:55:10.940730 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.440717426 +0000 UTC m=+157.136677978 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.041891 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.042268 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.542234964 +0000 UTC m=+157.238195516 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.042779 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.043109 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.543096469 +0000 UTC m=+157.239057021 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.144878 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.145423 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.64540472 +0000 UTC m=+157.341365272 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.246066 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.246473 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.746455965 +0000 UTC m=+157.442416517 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.347806 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.348066 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.848048315 +0000 UTC m=+157.544008867 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.410404 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" event={"ID":"5e23939f-1541-494e-af81-1cf62ff004b0","Type":"ContainerStarted","Data":"32c12ee6c3da3f0dcb8b96ff6a26efaa42882acbc713e171a32a7e506bd74376"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.413204 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" event={"ID":"e089d4b5-58b6-41ed-9901-4d01c86024ab","Type":"ContainerStarted","Data":"1c22f3e9a97d3df3efb1414d60ec8a619569fa0202cf05fc823711f28874cad5"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.414424 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" event={"ID":"956cc1d7-7e81-4e53-bae7-4adfe7139695","Type":"ContainerStarted","Data":"12bb0f179eb195196c90f5d6488736b6393fa4acff974449018ca19d17e8c760"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.416319 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vkf2v" event={"ID":"218bbc11-4dea-42a8-ab68-2ff2c2c84086","Type":"ContainerStarted","Data":"5e926f94850bbeef5585f8f1bd287a38d40aaebc15fb77d5bc5180e521a3c73b"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.416348 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-vkf2v" event={"ID":"218bbc11-4dea-42a8-ab68-2ff2c2c84086","Type":"ContainerStarted","Data":"139a96eed1819ce898b0add8895a1eaf2fb27bde4a12439edf54f6029c06dccf"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.440232 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.442622 4956 generic.go:334] "Generic (PLEG): container finished" podID="d2639202-023f-40ba-ad6e-291bc051dabe" containerID="0b47c13d2907c7dc393d82565da781ef0ea33cfa6c444f902cf2534ef72db73f" exitCode=0 Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.442789 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" event={"ID":"d2639202-023f-40ba-ad6e-291bc051dabe","Type":"ContainerDied","Data":"0b47c13d2907c7dc393d82565da781ef0ea33cfa6c444f902cf2534ef72db73f"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.451948 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.453191 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:11.953174617 +0000 UTC m=+157.649135169 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.463862 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-sb68p" event={"ID":"68745ad4-99d4-4028-95fb-0d70e3fba519","Type":"ContainerStarted","Data":"8938609d6ba946c9790ef411093726f2dffb890c53e48da0ac028e1a917b3fba"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.465220 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" event={"ID":"bb106cf7-e936-4502-a3bf-d154be0d6881","Type":"ContainerStarted","Data":"a7ec826526d9c1ec5dd99a3d32c15ab3435831dac6881a8f8bcea1d7d4d01e57"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.466591 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.467700 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" event={"ID":"e207f660-e7ee-4ad4-abe2-c2037c1af333","Type":"ContainerStarted","Data":"c5f931e00b4d5ea815b1b12972f22e4060b02531dc9a0c8d2a29a0655d0be7d9"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.467797 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" event={"ID":"e207f660-e7ee-4ad4-abe2-c2037c1af333","Type":"ContainerStarted","Data":"30eec9e377e01ceb3db00856fe9ecc3c49be9ec80139eaa820c588f351df9f08"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.468403 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.469000 4956 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-6gfsk container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" start-of-body= Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.469111 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" podUID="bb106cf7-e936-4502-a3bf-d154be0d6881" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.42:5443/healthz\": dial tcp 10.217.0.42:5443: connect: connection refused" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.471638 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" event={"ID":"08ce1e4a-c17a-4449-8cde-5febc97c501a","Type":"ContainerStarted","Data":"138d91c25370576f2b14dab3d65665e0176fd49fe2e6f1e1022804c821d7dfc9"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.471736 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" event={"ID":"08ce1e4a-c17a-4449-8cde-5febc97c501a","Type":"ContainerStarted","Data":"f785eb79134dca5a555ef077ca595291111c9a035d5a9b5f128b7a6adadb3401"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.473293 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" event={"ID":"4cf5e708-ff2f-4b3c-9653-1aace2d4bfff","Type":"ContainerStarted","Data":"204c0415476eec9a17ceed2e6985eba885da271f1791968673466f1996ca9c8a"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.477609 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" event={"ID":"183ca9fa-6509-45fb-8f0a-41161814ecbb","Type":"ContainerStarted","Data":"f2bdb37fcb5f5fbfd5a7e5ff7aa7826c4aab52e232de21a575e92c1f7ee7be7d"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.479834 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" event={"ID":"00806805-c6d3-4b27-aff6-3a6f6d03a094","Type":"ContainerStarted","Data":"4ad31ffd9d5b55e37b6db70fa4f4b3a6b34fc17a033e75680be235fda6883e0b"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.481522 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" event={"ID":"47afd50c-ef19-48a3-a7ad-98d6141719ec","Type":"ContainerStarted","Data":"85346ea1e6c613f9b33a533da0094a8b0cab3e229e8f2bbf115e140c7da6257c"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.482312 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.484243 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" event={"ID":"949267c7-0029-430b-a356-7059518c503a","Type":"ContainerStarted","Data":"ac572a8606ef582f221f45574379a8541d858701bf6fc226826c1750ce96ca8f"} Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.485029 4956 patch_prober.go:28] interesting pod/olm-operator-6b444d44fb-b9vfq container/olm-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" start-of-body= Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.485090 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" podUID="47afd50c-ef19-48a3-a7ad-98d6141719ec" containerName="olm-operator" probeResult="failure" output="Get \"https://10.217.0.40:8443/healthz\": dial tcp 10.217.0.40:8443: connect: connection refused" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.488971 4956 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bwzxf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.489013 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.530556 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" podStartSLOduration=131.5305291 podStartE2EDuration="2m11.5305291s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.51626084 +0000 UTC m=+157.212221402" watchObservedRunningTime="2025-11-26 16:55:11.5305291 +0000 UTC m=+157.226489652" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.553441 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.553718 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.053691146 +0000 UTC m=+157.749651708 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.556743 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.557480 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.057464615 +0000 UTC m=+157.753425167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.579326 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-zqnvj" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.609298 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" podStartSLOduration=131.609276474 podStartE2EDuration="2m11.609276474s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.601419828 +0000 UTC m=+157.297380390" watchObservedRunningTime="2025-11-26 16:55:11.609276474 +0000 UTC m=+157.305237036" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.648594 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" podStartSLOduration=131.648572913 podStartE2EDuration="2m11.648572913s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.646907216 +0000 UTC m=+157.342867788" watchObservedRunningTime="2025-11-26 16:55:11.648572913 +0000 UTC m=+157.344533465" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.664631 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-7txdb" podStartSLOduration=131.664609764 podStartE2EDuration="2m11.664609764s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.663203594 +0000 UTC m=+157.359164146" watchObservedRunningTime="2025-11-26 16:55:11.664609764 +0000 UTC m=+157.360570316" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.665636 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.665796 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.165776408 +0000 UTC m=+157.861736970 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.665964 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.666978 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.166969042 +0000 UTC m=+157.862929594 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.702896 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-shfrr" podStartSLOduration=131.702860014 podStartE2EDuration="2m11.702860014s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.700175677 +0000 UTC m=+157.396136229" watchObservedRunningTime="2025-11-26 16:55:11.702860014 +0000 UTC m=+157.398820556" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.734635 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:11 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:11 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:11 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.734726 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.761592 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" podStartSLOduration=131.761555351 podStartE2EDuration="2m11.761555351s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.756146466 +0000 UTC m=+157.452107018" watchObservedRunningTime="2025-11-26 16:55:11.761555351 +0000 UTC m=+157.457515903" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.767843 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.768205 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.268177582 +0000 UTC m=+157.964138124 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.839913 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-hs7zl" podStartSLOduration=131.839891043 podStartE2EDuration="2m11.839891043s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.838066651 +0000 UTC m=+157.534027213" watchObservedRunningTime="2025-11-26 16:55:11.839891043 +0000 UTC m=+157.535851595" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.841050 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" podStartSLOduration=131.841044396 podStartE2EDuration="2m11.841044396s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.806205055 +0000 UTC m=+157.502165617" watchObservedRunningTime="2025-11-26 16:55:11.841044396 +0000 UTC m=+157.537004948" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.869443 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.869900 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.369884615 +0000 UTC m=+158.065845167 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.897336 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-gkbcz" podStartSLOduration=131.897314784 podStartE2EDuration="2m11.897314784s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.894924955 +0000 UTC m=+157.590885507" watchObservedRunningTime="2025-11-26 16:55:11.897314784 +0000 UTC m=+157.593275336" Nov 26 16:55:11 crc kubenswrapper[4956]: I1126 16:55:11.970522 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:11 crc kubenswrapper[4956]: E1126 16:55:11.970948 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.47092539 +0000 UTC m=+158.166885942 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.005754 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-vkf2v" podStartSLOduration=8.00573699 podStartE2EDuration="8.00573699s" podCreationTimestamp="2025-11-26 16:55:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:11.965978197 +0000 UTC m=+157.661938749" watchObservedRunningTime="2025-11-26 16:55:12.00573699 +0000 UTC m=+157.701697542" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.006961 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9wvt8" podStartSLOduration=132.006953545 podStartE2EDuration="2m12.006953545s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:12.004112354 +0000 UTC m=+157.700072906" watchObservedRunningTime="2025-11-26 16:55:12.006953545 +0000 UTC m=+157.702914097" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.072745 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.073073 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.573059015 +0000 UTC m=+158.269019567 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.170340 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-p74bs"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.173684 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.177450 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.178272 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.178901 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.678865137 +0000 UTC m=+158.374825689 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.212009 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p74bs"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.279731 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gsv82\" (UniqueName: \"kubernetes.io/projected/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-kube-api-access-gsv82\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.279810 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.279833 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-catalog-content\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.279903 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-utilities\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.280236 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.78022141 +0000 UTC m=+158.476181962 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.345184 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-hsg7t"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.346609 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.356332 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.370352 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hsg7t"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.385694 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386040 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wckpr\" (UniqueName: \"kubernetes.io/projected/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-kube-api-access-wckpr\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386105 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-utilities\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386137 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gsv82\" (UniqueName: \"kubernetes.io/projected/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-kube-api-access-gsv82\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386166 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-catalog-content\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386185 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-utilities\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386226 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-catalog-content\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.386669 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-catalog-content\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.386765 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.886745442 +0000 UTC m=+158.582705994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.387004 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-utilities\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.425949 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gsv82\" (UniqueName: \"kubernetes.io/projected/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-kube-api-access-gsv82\") pod \"certified-operators-p74bs\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.487270 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-catalog-content\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.487313 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-utilities\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.487350 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.487377 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wckpr\" (UniqueName: \"kubernetes.io/projected/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-kube-api-access-wckpr\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.488112 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-catalog-content\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.488331 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-utilities\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.488588 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:12.98857743 +0000 UTC m=+158.684537982 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.529400 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" event={"ID":"e089d4b5-58b6-41ed-9901-4d01c86024ab","Type":"ContainerStarted","Data":"c2e3f6e396d430197f95aa43bc8fb70e6cc5cadc54f01fbe95a033a51c1ef9a2"} Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.530371 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.531197 4956 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bwzxf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.531245 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.588196 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.589722 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.089704556 +0000 UTC m=+158.785665108 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.605584 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-b9vfq" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.606631 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wckpr\" (UniqueName: \"kubernetes.io/projected/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-kube-api-access-wckpr\") pod \"community-operators-hsg7t\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.610174 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6ks5k"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.611723 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.643796 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ks5k"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.690209 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.691672 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-catalog-content\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.691725 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzmgz\" (UniqueName: \"kubernetes.io/projected/3540205e-79fd-4756-8203-d9b06a0ccf12-kube-api-access-rzmgz\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.691748 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-utilities\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.691822 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.692180 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.192165982 +0000 UTC m=+158.888126524 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.756278 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:12 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:12 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:12 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.756370 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.782087 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-cs6v7"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.783339 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.793954 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.794204 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-catalog-content\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.794250 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzmgz\" (UniqueName: \"kubernetes.io/projected/3540205e-79fd-4756-8203-d9b06a0ccf12-kube-api-access-rzmgz\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.794274 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-utilities\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.794729 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-utilities\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.794804 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.294786431 +0000 UTC m=+158.990746983 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.795040 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-catalog-content\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.813791 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cs6v7"] Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.843159 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzmgz\" (UniqueName: \"kubernetes.io/projected/3540205e-79fd-4756-8203-d9b06a0ccf12-kube-api-access-rzmgz\") pod \"certified-operators-6ks5k\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.899897 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-utilities\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.899971 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lkcgt\" (UniqueName: \"kubernetes.io/projected/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-kube-api-access-lkcgt\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.900005 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-catalog-content\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.900056 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:12 crc kubenswrapper[4956]: E1126 16:55:12.900419 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.400404678 +0000 UTC m=+159.096365220 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:12 crc kubenswrapper[4956]: I1126 16:55:12.951412 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.001803 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.002605 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-utilities\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.002879 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lkcgt\" (UniqueName: \"kubernetes.io/projected/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-kube-api-access-lkcgt\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.002923 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-catalog-content\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.003418 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-catalog-content\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.003509 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.503491001 +0000 UTC m=+159.199451553 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.003731 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-utilities\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.052393 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lkcgt\" (UniqueName: \"kubernetes.io/projected/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-kube-api-access-lkcgt\") pod \"community-operators-cs6v7\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.105043 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.105403 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.605388529 +0000 UTC m=+159.301349081 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.154280 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.209686 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.216311 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.716277176 +0000 UTC m=+159.412237728 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.317988 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.318471 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.818454103 +0000 UTC m=+159.514414655 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.339050 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.356164 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-6gfsk" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.361180 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-p74bs"] Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.424437 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6vqw5\" (UniqueName: \"kubernetes.io/projected/d2639202-023f-40ba-ad6e-291bc051dabe-kube-api-access-6vqw5\") pod \"d2639202-023f-40ba-ad6e-291bc051dabe\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.424653 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.424696 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2639202-023f-40ba-ad6e-291bc051dabe-config-volume\") pod \"d2639202-023f-40ba-ad6e-291bc051dabe\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.424728 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2639202-023f-40ba-ad6e-291bc051dabe-secret-volume\") pod \"d2639202-023f-40ba-ad6e-291bc051dabe\" (UID: \"d2639202-023f-40ba-ad6e-291bc051dabe\") " Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.424931 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.924891243 +0000 UTC m=+159.620851795 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.424989 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.425406 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:13.925399438 +0000 UTC m=+159.621359990 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.425597 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2639202-023f-40ba-ad6e-291bc051dabe-config-volume" (OuterVolumeSpecName: "config-volume") pod "d2639202-023f-40ba-ad6e-291bc051dabe" (UID: "d2639202-023f-40ba-ad6e-291bc051dabe"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.439905 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d2639202-023f-40ba-ad6e-291bc051dabe-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "d2639202-023f-40ba-ad6e-291bc051dabe" (UID: "d2639202-023f-40ba-ad6e-291bc051dabe"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.445199 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2639202-023f-40ba-ad6e-291bc051dabe-kube-api-access-6vqw5" (OuterVolumeSpecName: "kube-api-access-6vqw5") pod "d2639202-023f-40ba-ad6e-291bc051dabe" (UID: "d2639202-023f-40ba-ad6e-291bc051dabe"). InnerVolumeSpecName "kube-api-access-6vqw5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.529985 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.530273 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.030254432 +0000 UTC m=+159.726214984 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.530311 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.530373 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6vqw5\" (UniqueName: \"kubernetes.io/projected/d2639202-023f-40ba-ad6e-291bc051dabe-kube-api-access-6vqw5\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.530384 4956 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/d2639202-023f-40ba-ad6e-291bc051dabe-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.530395 4956 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/d2639202-023f-40ba-ad6e-291bc051dabe-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.530735 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.030727105 +0000 UTC m=+159.726687647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.559905 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p74bs" event={"ID":"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da","Type":"ContainerStarted","Data":"bba620648aed64b10b78eee8a3d0e7f4dc00d8988efb1802b3bffcc8f04cbdff"} Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.585042 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" event={"ID":"e089d4b5-58b6-41ed-9901-4d01c86024ab","Type":"ContainerStarted","Data":"88cf9d74d67a11239429ec3373fab57163fc8783c6b2c0b17a01b5606e4bf84a"} Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.587403 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-hsg7t"] Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.592267 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.592313 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402925-72d2x" event={"ID":"d2639202-023f-40ba-ad6e-291bc051dabe","Type":"ContainerDied","Data":"146167b02e819d75cf94b8c1b3dacc13296597bdce70399513d1a8c375934ab7"} Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.592379 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="146167b02e819d75cf94b8c1b3dacc13296597bdce70399513d1a8c375934ab7" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.631421 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.632360 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.132310435 +0000 UTC m=+159.828270987 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: W1126 16:55:13.682302 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda586fcfa_6e5e_40ff_b8b7_bc2f26f0b732.slice/crio-6dc1d3bdf8c56e1b6c0f73d9dd78d456328faf723b6dca66fc1a5a3f6af8eec1 WatchSource:0}: Error finding container 6dc1d3bdf8c56e1b6c0f73d9dd78d456328faf723b6dca66fc1a5a3f6af8eec1: Status 404 returned error can't find the container with id 6dc1d3bdf8c56e1b6c0f73d9dd78d456328faf723b6dca66fc1a5a3f6af8eec1 Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.734147 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.734589 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.234574765 +0000 UTC m=+159.930535317 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.737480 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:13 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:13 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:13 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.737547 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.764578 4956 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.795933 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6ks5k"] Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.837310 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.838156 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.33808488 +0000 UTC m=+160.034045432 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:13 crc kubenswrapper[4956]: W1126 16:55:13.859376 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3540205e_79fd_4756_8203_d9b06a0ccf12.slice/crio-992fd210af8b1325c3b5f4a501a1d4d8be8746c9ea23d2763eb7891cf57e5df7 WatchSource:0}: Error finding container 992fd210af8b1325c3b5f4a501a1d4d8be8746c9ea23d2763eb7891cf57e5df7: Status 404 returned error can't find the container with id 992fd210af8b1325c3b5f4a501a1d4d8be8746c9ea23d2763eb7891cf57e5df7 Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.886410 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-cs6v7"] Nov 26 16:55:13 crc kubenswrapper[4956]: I1126 16:55:13.939130 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:13 crc kubenswrapper[4956]: E1126 16:55:13.939671 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.43965386 +0000 UTC m=+160.135614412 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.041033 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:14 crc kubenswrapper[4956]: E1126 16:55:14.041220 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.541194099 +0000 UTC m=+160.237154651 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.041354 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:14 crc kubenswrapper[4956]: E1126 16:55:14.041762 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.541751095 +0000 UTC m=+160.237711647 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.148035 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:14 crc kubenswrapper[4956]: E1126 16:55:14.148312 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.648273207 +0000 UTC m=+160.344233759 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.148435 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:14 crc kubenswrapper[4956]: E1126 16:55:14.148906 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-11-26 16:55:14.648886274 +0000 UTC m=+160.344846826 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-xjxmw" (UID: "8394076a-e17a-4cea-a904-eb269c237f00") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.163853 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sj"] Nov 26 16:55:14 crc kubenswrapper[4956]: E1126 16:55:14.164139 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2639202-023f-40ba-ad6e-291bc051dabe" containerName="collect-profiles" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.164156 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2639202-023f-40ba-ad6e-291bc051dabe" containerName="collect-profiles" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.164291 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2639202-023f-40ba-ad6e-291bc051dabe" containerName="collect-profiles" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.165174 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.168420 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.176997 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sj"] Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.200272 4956 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-11-26T16:55:13.764630959Z","Handler":null,"Name":""} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.203728 4956 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.203759 4956 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.250165 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.250473 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-utilities\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.250569 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-59dhh\" (UniqueName: \"kubernetes.io/projected/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-kube-api-access-59dhh\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.250611 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-catalog-content\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.257753 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.352168 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-utilities\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.352761 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.353031 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-utilities\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.354192 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-59dhh\" (UniqueName: \"kubernetes.io/projected/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-kube-api-access-59dhh\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.354230 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-catalog-content\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.354966 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-catalog-content\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.356450 4956 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.356489 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.379992 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-59dhh\" (UniqueName: \"kubernetes.io/projected/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-kube-api-access-59dhh\") pod \"redhat-marketplace-6z8sj\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.405808 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-xjxmw\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.484388 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.519248 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.539349 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-q2bxq"] Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.540444 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.555155 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2bxq"] Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.599773 4956 generic.go:334] "Generic (PLEG): container finished" podID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerID="9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b" exitCode=0 Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.599923 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerDied","Data":"9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.599977 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerStarted","Data":"2361cfcdc662d59e2d967b8f17800be51dd17ec669bb33314a0ba8c48661ed89"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.602243 4956 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.618314 4956 generic.go:334] "Generic (PLEG): container finished" podID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerID="dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8" exitCode=0 Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.618486 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerDied","Data":"dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.618539 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerStarted","Data":"6dc1d3bdf8c56e1b6c0f73d9dd78d456328faf723b6dca66fc1a5a3f6af8eec1"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.622906 4956 generic.go:334] "Generic (PLEG): container finished" podID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerID="9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40" exitCode=0 Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.622955 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ks5k" event={"ID":"3540205e-79fd-4756-8203-d9b06a0ccf12","Type":"ContainerDied","Data":"9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.622976 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ks5k" event={"ID":"3540205e-79fd-4756-8203-d9b06a0ccf12","Type":"ContainerStarted","Data":"992fd210af8b1325c3b5f4a501a1d4d8be8746c9ea23d2763eb7891cf57e5df7"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.631391 4956 generic.go:334] "Generic (PLEG): container finished" podID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerID="e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245" exitCode=0 Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.631507 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p74bs" event={"ID":"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da","Type":"ContainerDied","Data":"e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.647758 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" event={"ID":"e089d4b5-58b6-41ed-9901-4d01c86024ab","Type":"ContainerStarted","Data":"9144edf2ae592fba6c018ab9d284d998854bb22c5d49eb8501126cbb2dcd3a1c"} Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.672662 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-utilities\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.672745 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-54jhp\" (UniqueName: \"kubernetes.io/projected/2fafdf53-76b4-49ba-8346-430e2bd5d345-kube-api-access-54jhp\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.672804 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-catalog-content\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.738738 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:14 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:14 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:14 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.739272 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.760219 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-8htxc" podStartSLOduration=11.760189846 podStartE2EDuration="11.760189846s" podCreationTimestamp="2025-11-26 16:55:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:14.753073742 +0000 UTC m=+160.449034304" watchObservedRunningTime="2025-11-26 16:55:14.760189846 +0000 UTC m=+160.456150388" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.775035 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-catalog-content\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.775177 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-utilities\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.775259 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-54jhp\" (UniqueName: \"kubernetes.io/projected/2fafdf53-76b4-49ba-8346-430e2bd5d345-kube-api-access-54jhp\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.777492 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-utilities\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.777609 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-catalog-content\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.801568 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-54jhp\" (UniqueName: \"kubernetes.io/projected/2fafdf53-76b4-49ba-8346-430e2bd5d345-kube-api-access-54jhp\") pod \"redhat-marketplace-q2bxq\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.874022 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sj"] Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.874914 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:55:14 crc kubenswrapper[4956]: I1126 16:55:14.982621 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xjxmw"] Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.011531 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.110230 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2bxq"] Nov 26 16:55:15 crc kubenswrapper[4956]: W1126 16:55:15.148018 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2fafdf53_76b4_49ba_8346_430e2bd5d345.slice/crio-63cf40b760a44d91683e1aa3cdfd56ade9de67a6a5e0898de7554b183fa6ddd9 WatchSource:0}: Error finding container 63cf40b760a44d91683e1aa3cdfd56ade9de67a6a5e0898de7554b183fa6ddd9: Status 404 returned error can't find the container with id 63cf40b760a44d91683e1aa3cdfd56ade9de67a6a5e0898de7554b183fa6ddd9 Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.262380 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.262459 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.275888 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.364399 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.372415 4956 patch_prober.go:28] interesting pod/downloads-7954f5f757-cwlp6 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.372515 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-cwlp6" podUID="996f361a-f4aa-4d79-b12c-3e357f126f43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.372549 4956 patch_prober.go:28] interesting pod/downloads-7954f5f757-cwlp6 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" start-of-body= Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.372600 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-cwlp6" podUID="996f361a-f4aa-4d79-b12c-3e357f126f43" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.17:8080/\": dial tcp 10.217.0.17:8080: connect: connection refused" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.477496 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.508185 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.508253 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.510443 4956 patch_prober.go:28] interesting pod/console-f9d7485db-ssdwk container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" start-of-body= Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.510550 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-ssdwk" podUID="166d8e6f-9af9-4212-9593-0b47d27918d1" containerName="console" probeResult="failure" output="Get \"https://10.217.0.8:8443/health\": dial tcp 10.217.0.8:8443: connect: connection refused" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.538358 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zc7rm"] Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.540379 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.541359 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.541752 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.560823 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.572681 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.583397 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zc7rm"] Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.605749 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-nd5x4" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.670217 4956 generic.go:334] "Generic (PLEG): container finished" podID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerID="9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210" exitCode=0 Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.671202 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2bxq" event={"ID":"2fafdf53-76b4-49ba-8346-430e2bd5d345","Type":"ContainerDied","Data":"9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210"} Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.671243 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2bxq" event={"ID":"2fafdf53-76b4-49ba-8346-430e2bd5d345","Type":"ContainerStarted","Data":"63cf40b760a44d91683e1aa3cdfd56ade9de67a6a5e0898de7554b183fa6ddd9"} Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.681456 4956 generic.go:334] "Generic (PLEG): container finished" podID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerID="241647a8d8ad17f8dbd7f8adbcc13858bf96e43e3a7ac174a29bbd741500d3ad" exitCode=0 Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.681580 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sj" event={"ID":"9a71e654-5284-4cbb-a3de-59b7e29d1fc9","Type":"ContainerDied","Data":"241647a8d8ad17f8dbd7f8adbcc13858bf96e43e3a7ac174a29bbd741500d3ad"} Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.681611 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sj" event={"ID":"9a71e654-5284-4cbb-a3de-59b7e29d1fc9","Type":"ContainerStarted","Data":"8c4e6e54fadd1c1e5f40650a7445eff3fea2898524334d102e83b297e11334a3"} Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.685639 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" event={"ID":"8394076a-e17a-4cea-a904-eb269c237f00","Type":"ContainerStarted","Data":"cd18ff7527c1e635622e62921b7abd26b075602ccdb89553a273b90078208974"} Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.685677 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.685688 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" event={"ID":"8394076a-e17a-4cea-a904-eb269c237f00","Type":"ContainerStarted","Data":"d328275e411872b8ef64c2a4e33fc2eec646ca0c742824ea6bd3b974b4338bd2"} Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.693419 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-catalog-content\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.693579 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-utilities\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.693624 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6jd9\" (UniqueName: \"kubernetes.io/projected/8aa59c7b-5919-4987-b0d4-cc4a629ee006-kube-api-access-g6jd9\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.695154 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-fx7df" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.698021 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-f2ldn" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.736034 4956 patch_prober.go:28] interesting pod/router-default-5444994796-r245h container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Nov 26 16:55:15 crc kubenswrapper[4956]: [-]has-synced failed: reason withheld Nov 26 16:55:15 crc kubenswrapper[4956]: [+]process-running ok Nov 26 16:55:15 crc kubenswrapper[4956]: healthz check failed Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.736413 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-r245h" podUID="ffef519e-8b2e-4a79-b15b-a6c3bb36a875" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.766466 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" podStartSLOduration=135.766439461 podStartE2EDuration="2m15.766439461s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:15.766433381 +0000 UTC m=+161.462393943" watchObservedRunningTime="2025-11-26 16:55:15.766439461 +0000 UTC m=+161.462400013" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.795076 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-utilities\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.795156 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6jd9\" (UniqueName: \"kubernetes.io/projected/8aa59c7b-5919-4987-b0d4-cc4a629ee006-kube-api-access-g6jd9\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.795245 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-catalog-content\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.796345 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-utilities\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.798767 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-catalog-content\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.856228 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6jd9\" (UniqueName: \"kubernetes.io/projected/8aa59c7b-5919-4987-b0d4-cc4a629ee006-kube-api-access-g6jd9\") pod \"redhat-operators-zc7rm\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.881588 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.990700 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-89rgc"] Nov 26 16:55:15 crc kubenswrapper[4956]: I1126 16:55:15.991857 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.035698 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-89rgc"] Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.101444 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-catalog-content\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.101536 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6kms5\" (UniqueName: \"kubernetes.io/projected/638a6242-d034-4b5d-90f9-f25fcf1bc742-kube-api-access-6kms5\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.101557 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-utilities\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.202832 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6kms5\" (UniqueName: \"kubernetes.io/projected/638a6242-d034-4b5d-90f9-f25fcf1bc742-kube-api-access-6kms5\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.202961 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-utilities\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.203091 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-catalog-content\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.203885 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-catalog-content\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.207163 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-utilities\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.248424 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6kms5\" (UniqueName: \"kubernetes.io/projected/638a6242-d034-4b5d-90f9-f25fcf1bc742-kube-api-access-6kms5\") pod \"redhat-operators-89rgc\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.329355 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.367048 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zc7rm"] Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.438444 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.439549 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.470543 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.477860 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Nov 26 16:55:16 crc kubenswrapper[4956]: W1126 16:55:16.478185 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8aa59c7b_5919_4987_b0d4_cc4a629ee006.slice/crio-3747b156728be566238564e674ba49104ae4337d57b13baedf4742ed09afac42 WatchSource:0}: Error finding container 3747b156728be566238564e674ba49104ae4337d57b13baedf4742ed09afac42: Status 404 returned error can't find the container with id 3747b156728be566238564e674ba49104ae4337d57b13baedf4742ed09afac42 Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.495293 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.511312 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da69b41a-6323-4098-830b-e0745a21adc6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.511357 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da69b41a-6323-4098-830b-e0745a21adc6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.612267 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da69b41a-6323-4098-830b-e0745a21adc6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.612602 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da69b41a-6323-4098-830b-e0745a21adc6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.612426 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da69b41a-6323-4098-830b-e0745a21adc6-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.645810 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da69b41a-6323-4098-830b-e0745a21adc6-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.702378 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerStarted","Data":"3747b156728be566238564e674ba49104ae4337d57b13baedf4742ed09afac42"} Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.727457 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.731775 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.748541 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.802589 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:16 crc kubenswrapper[4956]: I1126 16:55:16.923051 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-89rgc"] Nov 26 16:55:16 crc kubenswrapper[4956]: W1126 16:55:16.954680 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod638a6242_d034_4b5d_90f9_f25fcf1bc742.slice/crio-cc097537545aef922936d70eec2011725e24ff5a3713df41a9fed6410bf4dc16 WatchSource:0}: Error finding container cc097537545aef922936d70eec2011725e24ff5a3713df41a9fed6410bf4dc16: Status 404 returned error can't find the container with id cc097537545aef922936d70eec2011725e24ff5a3713df41a9fed6410bf4dc16 Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.422265 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.720454 4956 generic.go:334] "Generic (PLEG): container finished" podID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerID="1fe2cfe8cfbfd5ce0e736a894a46405247b47b744a903aaa198283a2d1ade092" exitCode=0 Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.720959 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerDied","Data":"1fe2cfe8cfbfd5ce0e736a894a46405247b47b744a903aaa198283a2d1ade092"} Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.731207 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"da69b41a-6323-4098-830b-e0745a21adc6","Type":"ContainerStarted","Data":"7ef4618f6371ce5c329fce9547d273077ba3863a8e6fe4b9c053e23a7ac34ff0"} Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.738837 4956 generic.go:334] "Generic (PLEG): container finished" podID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerID="6d1cd7d539adde2b27026f8a3021441a115f9738c68e0e8f36068c3e1ddb8d46" exitCode=0 Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.740042 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89rgc" event={"ID":"638a6242-d034-4b5d-90f9-f25fcf1bc742","Type":"ContainerDied","Data":"6d1cd7d539adde2b27026f8a3021441a115f9738c68e0e8f36068c3e1ddb8d46"} Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.740150 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89rgc" event={"ID":"638a6242-d034-4b5d-90f9-f25fcf1bc742","Type":"ContainerStarted","Data":"cc097537545aef922936d70eec2011725e24ff5a3713df41a9fed6410bf4dc16"} Nov 26 16:55:17 crc kubenswrapper[4956]: I1126 16:55:17.745903 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-r245h" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.777893 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"da69b41a-6323-4098-830b-e0745a21adc6","Type":"ContainerStarted","Data":"95e97c8ff14e0ed9e870e967ebe53959ad79dfbff376084607992d96501572ad"} Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.795372 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/revision-pruner-9-crc" podStartSLOduration=2.795346456 podStartE2EDuration="2.795346456s" podCreationTimestamp="2025-11-26 16:55:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:18.793568255 +0000 UTC m=+164.489528817" watchObservedRunningTime="2025-11-26 16:55:18.795346456 +0000 UTC m=+164.491307008" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.820230 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.821240 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.838156 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.838349 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.862831 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.862985 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.881883 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.966003 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.966384 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.966435 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:18 crc kubenswrapper[4956]: I1126 16:55:18.996415 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:19 crc kubenswrapper[4956]: I1126 16:55:19.161521 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:19 crc kubenswrapper[4956]: I1126 16:55:19.165548 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-vkf2v" Nov 26 16:55:19 crc kubenswrapper[4956]: I1126 16:55:19.803313 4956 generic.go:334] "Generic (PLEG): container finished" podID="da69b41a-6323-4098-830b-e0745a21adc6" containerID="95e97c8ff14e0ed9e870e967ebe53959ad79dfbff376084607992d96501572ad" exitCode=0 Nov 26 16:55:19 crc kubenswrapper[4956]: I1126 16:55:19.803367 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"da69b41a-6323-4098-830b-e0745a21adc6","Type":"ContainerDied","Data":"95e97c8ff14e0ed9e870e967ebe53959ad79dfbff376084607992d96501572ad"} Nov 26 16:55:19 crc kubenswrapper[4956]: I1126 16:55:19.898172 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Nov 26 16:55:20 crc kubenswrapper[4956]: I1126 16:55:20.834667 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c3360f5-aa2a-4848-9c3c-7a238f8a761f","Type":"ContainerStarted","Data":"6a936013eff949ac68200119ff4ee2f8410e073c952f53a39a65a5a88c98fffd"} Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.177343 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.233004 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da69b41a-6323-4098-830b-e0745a21adc6-kubelet-dir\") pod \"da69b41a-6323-4098-830b-e0745a21adc6\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.233677 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da69b41a-6323-4098-830b-e0745a21adc6-kube-api-access\") pod \"da69b41a-6323-4098-830b-e0745a21adc6\" (UID: \"da69b41a-6323-4098-830b-e0745a21adc6\") " Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.233119 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/da69b41a-6323-4098-830b-e0745a21adc6-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "da69b41a-6323-4098-830b-e0745a21adc6" (UID: "da69b41a-6323-4098-830b-e0745a21adc6"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.247159 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/da69b41a-6323-4098-830b-e0745a21adc6-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "da69b41a-6323-4098-830b-e0745a21adc6" (UID: "da69b41a-6323-4098-830b-e0745a21adc6"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.337555 4956 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/da69b41a-6323-4098-830b-e0745a21adc6-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.337633 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/da69b41a-6323-4098-830b-e0745a21adc6-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.853389 4956 generic.go:334] "Generic (PLEG): container finished" podID="7c3360f5-aa2a-4848-9c3c-7a238f8a761f" containerID="19003ccbb44b2294f57e04cc9f890ad6daa8a0d446bb02612b9930badbb91a0e" exitCode=0 Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.853492 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c3360f5-aa2a-4848-9c3c-7a238f8a761f","Type":"ContainerDied","Data":"19003ccbb44b2294f57e04cc9f890ad6daa8a0d446bb02612b9930badbb91a0e"} Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.858456 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"da69b41a-6323-4098-830b-e0745a21adc6","Type":"ContainerDied","Data":"7ef4618f6371ce5c329fce9547d273077ba3863a8e6fe4b9c053e23a7ac34ff0"} Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.858490 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7ef4618f6371ce5c329fce9547d273077ba3863a8e6fe4b9c053e23a7ac34ff0" Nov 26 16:55:21 crc kubenswrapper[4956]: I1126 16:55:21.858550 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Nov 26 16:55:22 crc kubenswrapper[4956]: I1126 16:55:22.352581 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 16:55:23 crc kubenswrapper[4956]: I1126 16:55:23.481788 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:55:23 crc kubenswrapper[4956]: I1126 16:55:23.487334 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0956e12e-96b8-4b83-bc0a-666639b36586-metrics-certs\") pod \"network-metrics-daemon-b8nt2\" (UID: \"0956e12e-96b8-4b83-bc0a-666639b36586\") " pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:55:23 crc kubenswrapper[4956]: I1126 16:55:23.617462 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-b8nt2" Nov 26 16:55:25 crc kubenswrapper[4956]: I1126 16:55:25.385897 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-cwlp6" Nov 26 16:55:25 crc kubenswrapper[4956]: I1126 16:55:25.529773 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:25 crc kubenswrapper[4956]: I1126 16:55:25.536137 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-ssdwk" Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.833835 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.921821 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"7c3360f5-aa2a-4848-9c3c-7a238f8a761f","Type":"ContainerDied","Data":"6a936013eff949ac68200119ff4ee2f8410e073c952f53a39a65a5a88c98fffd"} Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.921888 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6a936013eff949ac68200119ff4ee2f8410e073c952f53a39a65a5a88c98fffd" Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.921933 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.982046 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kube-api-access\") pod \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.982126 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kubelet-dir\") pod \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\" (UID: \"7c3360f5-aa2a-4848-9c3c-7a238f8a761f\") " Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.982353 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7c3360f5-aa2a-4848-9c3c-7a238f8a761f" (UID: "7c3360f5-aa2a-4848-9c3c-7a238f8a761f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 16:55:28 crc kubenswrapper[4956]: I1126 16:55:28.989699 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7c3360f5-aa2a-4848-9c3c-7a238f8a761f" (UID: "7c3360f5-aa2a-4848-9c3c-7a238f8a761f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:55:29 crc kubenswrapper[4956]: I1126 16:55:29.083581 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:29 crc kubenswrapper[4956]: I1126 16:55:29.083624 4956 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7c3360f5-aa2a-4848-9c3c-7a238f8a761f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 16:55:33 crc kubenswrapper[4956]: I1126 16:55:33.237153 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Nov 26 16:55:34 crc kubenswrapper[4956]: I1126 16:55:34.492488 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 16:55:39 crc kubenswrapper[4956]: I1126 16:55:39.550658 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 16:55:39 crc kubenswrapper[4956]: I1126 16:55:39.551397 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 16:55:46 crc kubenswrapper[4956]: E1126 16:55:46.119603 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-operator-index:v4.18" Nov 26 16:55:46 crc kubenswrapper[4956]: E1126 16:55:46.120547 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-g6jd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-operators-zc7rm_openshift-marketplace(8aa59c7b-5919-4987-b0d4-cc4a629ee006): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:46 crc kubenswrapper[4956]: E1126 16:55:46.122294 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-operators-zc7rm" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" Nov 26 16:55:47 crc kubenswrapper[4956]: I1126 16:55:47.095986 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-bztls" Nov 26 16:55:47 crc kubenswrapper[4956]: E1126 16:55:47.182564 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-operators-zc7rm" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" Nov 26 16:55:48 crc kubenswrapper[4956]: E1126 16:55:48.855904 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 26 16:55:48 crc kubenswrapper[4956]: E1126 16:55:48.856570 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-54jhp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-q2bxq_openshift-marketplace(2fafdf53-76b4-49ba-8346-430e2bd5d345): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:48 crc kubenswrapper[4956]: E1126 16:55:48.858031 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-q2bxq" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" Nov 26 16:55:50 crc kubenswrapper[4956]: E1126 16:55:50.128192 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-q2bxq" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" Nov 26 16:55:50 crc kubenswrapper[4956]: E1126 16:55:50.199971 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Nov 26 16:55:50 crc kubenswrapper[4956]: E1126 16:55:50.200169 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-59dhh,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-6z8sj_openshift-marketplace(9a71e654-5284-4cbb-a3de-59b7e29d1fc9): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:50 crc kubenswrapper[4956]: E1126 16:55:50.201459 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-6z8sj" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" Nov 26 16:55:52 crc kubenswrapper[4956]: E1126 16:55:52.389295 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 26 16:55:52 crc kubenswrapper[4956]: E1126 16:55:52.389881 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gsv82,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-p74bs_openshift-marketplace(5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:52 crc kubenswrapper[4956]: E1126 16:55:52.391321 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-p74bs" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" Nov 26 16:55:53 crc kubenswrapper[4956]: E1126 16:55:53.741143 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Nov 26 16:55:53 crc kubenswrapper[4956]: E1126 16:55:53.741329 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rzmgz,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-6ks5k_openshift-marketplace(3540205e-79fd-4756-8203-d9b06a0ccf12): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:53 crc kubenswrapper[4956]: E1126 16:55:53.742521 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-6ks5k" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" Nov 26 16:55:53 crc kubenswrapper[4956]: E1126 16:55:53.822166 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-6z8sj" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" Nov 26 16:55:53 crc kubenswrapper[4956]: E1126 16:55:53.822166 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-p74bs" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" Nov 26 16:55:54 crc kubenswrapper[4956]: E1126 16:55:54.094451 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-6ks5k" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" Nov 26 16:55:54 crc kubenswrapper[4956]: I1126 16:55:54.286286 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-b8nt2"] Nov 26 16:55:54 crc kubenswrapper[4956]: W1126 16:55:54.294711 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0956e12e_96b8_4b83_bc0a_666639b36586.slice/crio-a24a91d0d22f77243e88ad95282d62dd28ed8ebd8dfb504446e396343fb4ffde WatchSource:0}: Error finding container a24a91d0d22f77243e88ad95282d62dd28ed8ebd8dfb504446e396343fb4ffde: Status 404 returned error can't find the container with id a24a91d0d22f77243e88ad95282d62dd28ed8ebd8dfb504446e396343fb4ffde Nov 26 16:55:55 crc kubenswrapper[4956]: I1126 16:55:55.098179 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" event={"ID":"0956e12e-96b8-4b83-bc0a-666639b36586","Type":"ContainerStarted","Data":"cf69a4e6750d8945080c90af613db73a5a7614f749319021be33861fa41bbe61"} Nov 26 16:55:55 crc kubenswrapper[4956]: I1126 16:55:55.098249 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" event={"ID":"0956e12e-96b8-4b83-bc0a-666639b36586","Type":"ContainerStarted","Data":"a24a91d0d22f77243e88ad95282d62dd28ed8ebd8dfb504446e396343fb4ffde"} Nov 26 16:55:55 crc kubenswrapper[4956]: E1126 16:55:55.907885 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 26 16:55:55 crc kubenswrapper[4956]: E1126 16:55:55.908631 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-wckpr,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-hsg7t_openshift-marketplace(a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:55 crc kubenswrapper[4956]: E1126 16:55:55.909911 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-hsg7t" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" Nov 26 16:55:55 crc kubenswrapper[4956]: E1126 16:55:55.943473 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Nov 26 16:55:55 crc kubenswrapper[4956]: E1126 16:55:55.943663 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lkcgt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-cs6v7_openshift-marketplace(f1fdb8dd-363c-42b9-ab6e-862d8b459f90): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Nov 26 16:55:55 crc kubenswrapper[4956]: E1126 16:55:55.944843 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-cs6v7" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" Nov 26 16:55:56 crc kubenswrapper[4956]: E1126 16:55:56.103992 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-cs6v7" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" Nov 26 16:55:56 crc kubenswrapper[4956]: E1126 16:55:56.104407 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-hsg7t" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" Nov 26 16:55:57 crc kubenswrapper[4956]: I1126 16:55:57.111929 4956 generic.go:334] "Generic (PLEG): container finished" podID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerID="efb0f298147341ab4b0acf25ce1681c0f8a8eadc1f6bf8db610e00bebed5a7c6" exitCode=0 Nov 26 16:55:57 crc kubenswrapper[4956]: I1126 16:55:57.112050 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89rgc" event={"ID":"638a6242-d034-4b5d-90f9-f25fcf1bc742","Type":"ContainerDied","Data":"efb0f298147341ab4b0acf25ce1681c0f8a8eadc1f6bf8db610e00bebed5a7c6"} Nov 26 16:55:57 crc kubenswrapper[4956]: I1126 16:55:57.116643 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-b8nt2" event={"ID":"0956e12e-96b8-4b83-bc0a-666639b36586","Type":"ContainerStarted","Data":"fec323e96f85c4abc9860b9489b237c44f5a39b71a967d2836de618c02917264"} Nov 26 16:55:57 crc kubenswrapper[4956]: I1126 16:55:57.148146 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-b8nt2" podStartSLOduration=177.148125591 podStartE2EDuration="2m57.148125591s" podCreationTimestamp="2025-11-26 16:53:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:55:57.145577648 +0000 UTC m=+202.841538200" watchObservedRunningTime="2025-11-26 16:55:57.148125591 +0000 UTC m=+202.844086143" Nov 26 16:55:58 crc kubenswrapper[4956]: I1126 16:55:58.126057 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89rgc" event={"ID":"638a6242-d034-4b5d-90f9-f25fcf1bc742","Type":"ContainerStarted","Data":"3ed2b1c5ffe0f3394ca5d11fb4ee10a26bb1708f57f930ac908e9ba00bcce628"} Nov 26 16:55:58 crc kubenswrapper[4956]: I1126 16:55:58.148902 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-89rgc" podStartSLOduration=4.15624716 podStartE2EDuration="43.148879428s" podCreationTimestamp="2025-11-26 16:55:15 +0000 UTC" firstStartedPulling="2025-11-26 16:55:18.780719876 +0000 UTC m=+164.476680428" lastFinishedPulling="2025-11-26 16:55:57.773352144 +0000 UTC m=+203.469312696" observedRunningTime="2025-11-26 16:55:58.146457968 +0000 UTC m=+203.842418520" watchObservedRunningTime="2025-11-26 16:55:58.148879428 +0000 UTC m=+203.844839980" Nov 26 16:56:03 crc kubenswrapper[4956]: I1126 16:56:03.167094 4956 generic.go:334] "Generic (PLEG): container finished" podID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerID="a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526" exitCode=0 Nov 26 16:56:03 crc kubenswrapper[4956]: I1126 16:56:03.167204 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2bxq" event={"ID":"2fafdf53-76b4-49ba-8346-430e2bd5d345","Type":"ContainerDied","Data":"a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526"} Nov 26 16:56:03 crc kubenswrapper[4956]: I1126 16:56:03.170116 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerStarted","Data":"c843b0880e491a180db8b43d7cfc56ccfa6cf86c49b4bc95b98271546ffa1021"} Nov 26 16:56:04 crc kubenswrapper[4956]: I1126 16:56:04.180286 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2bxq" event={"ID":"2fafdf53-76b4-49ba-8346-430e2bd5d345","Type":"ContainerStarted","Data":"381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd"} Nov 26 16:56:04 crc kubenswrapper[4956]: I1126 16:56:04.183310 4956 generic.go:334] "Generic (PLEG): container finished" podID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerID="c843b0880e491a180db8b43d7cfc56ccfa6cf86c49b4bc95b98271546ffa1021" exitCode=0 Nov 26 16:56:04 crc kubenswrapper[4956]: I1126 16:56:04.183344 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerDied","Data":"c843b0880e491a180db8b43d7cfc56ccfa6cf86c49b4bc95b98271546ffa1021"} Nov 26 16:56:04 crc kubenswrapper[4956]: I1126 16:56:04.208224 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-q2bxq" podStartSLOduration=2.2540555270000002 podStartE2EDuration="50.208198582s" podCreationTimestamp="2025-11-26 16:55:14 +0000 UTC" firstStartedPulling="2025-11-26 16:55:15.6752704 +0000 UTC m=+161.371230952" lastFinishedPulling="2025-11-26 16:56:03.629413455 +0000 UTC m=+209.325374007" observedRunningTime="2025-11-26 16:56:04.206954616 +0000 UTC m=+209.902915188" watchObservedRunningTime="2025-11-26 16:56:04.208198582 +0000 UTC m=+209.904159134" Nov 26 16:56:04 crc kubenswrapper[4956]: I1126 16:56:04.875657 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:56:04 crc kubenswrapper[4956]: I1126 16:56:04.877112 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:56:05 crc kubenswrapper[4956]: I1126 16:56:05.201945 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerStarted","Data":"a5a3ef711b91179e96ea3b6f2bf744d56766cb21a99cf44f989b96e1dff6587c"} Nov 26 16:56:05 crc kubenswrapper[4956]: I1126 16:56:05.229354 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zc7rm" podStartSLOduration=3.306667169 podStartE2EDuration="50.229280883s" podCreationTimestamp="2025-11-26 16:55:15 +0000 UTC" firstStartedPulling="2025-11-26 16:55:17.72815054 +0000 UTC m=+163.424111092" lastFinishedPulling="2025-11-26 16:56:04.650764264 +0000 UTC m=+210.346724806" observedRunningTime="2025-11-26 16:56:05.226051211 +0000 UTC m=+210.922011773" watchObservedRunningTime="2025-11-26 16:56:05.229280883 +0000 UTC m=+210.925241435" Nov 26 16:56:05 crc kubenswrapper[4956]: I1126 16:56:05.436733 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8tmqv"] Nov 26 16:56:05 crc kubenswrapper[4956]: I1126 16:56:05.883265 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:56:05 crc kubenswrapper[4956]: I1126 16:56:05.883331 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:56:06 crc kubenswrapper[4956]: I1126 16:56:06.065541 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-marketplace-q2bxq" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="registry-server" probeResult="failure" output=< Nov 26 16:56:06 crc kubenswrapper[4956]: timeout: failed to connect service ":50051" within 1s Nov 26 16:56:06 crc kubenswrapper[4956]: > Nov 26 16:56:06 crc kubenswrapper[4956]: I1126 16:56:06.330489 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:56:06 crc kubenswrapper[4956]: I1126 16:56:06.330702 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:56:06 crc kubenswrapper[4956]: I1126 16:56:06.392527 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:56:06 crc kubenswrapper[4956]: I1126 16:56:06.942512 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-zc7rm" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="registry-server" probeResult="failure" output=< Nov 26 16:56:06 crc kubenswrapper[4956]: timeout: failed to connect service ":50051" within 1s Nov 26 16:56:06 crc kubenswrapper[4956]: > Nov 26 16:56:07 crc kubenswrapper[4956]: I1126 16:56:07.202496 4956 generic.go:334] "Generic (PLEG): container finished" podID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerID="6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2" exitCode=0 Nov 26 16:56:07 crc kubenswrapper[4956]: I1126 16:56:07.202581 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ks5k" event={"ID":"3540205e-79fd-4756-8203-d9b06a0ccf12","Type":"ContainerDied","Data":"6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2"} Nov 26 16:56:07 crc kubenswrapper[4956]: I1126 16:56:07.206595 4956 generic.go:334] "Generic (PLEG): container finished" podID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerID="1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec" exitCode=0 Nov 26 16:56:07 crc kubenswrapper[4956]: I1126 16:56:07.207027 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p74bs" event={"ID":"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da","Type":"ContainerDied","Data":"1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec"} Nov 26 16:56:07 crc kubenswrapper[4956]: I1126 16:56:07.276767 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:56:09 crc kubenswrapper[4956]: I1126 16:56:09.550964 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 16:56:09 crc kubenswrapper[4956]: I1126 16:56:09.551485 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 16:56:09 crc kubenswrapper[4956]: I1126 16:56:09.551558 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:56:09 crc kubenswrapper[4956]: I1126 16:56:09.552442 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 16:56:09 crc kubenswrapper[4956]: I1126 16:56:09.552656 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf" gracePeriod=600 Nov 26 16:56:10 crc kubenswrapper[4956]: I1126 16:56:10.221703 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf" exitCode=0 Nov 26 16:56:10 crc kubenswrapper[4956]: I1126 16:56:10.221784 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf"} Nov 26 16:56:11 crc kubenswrapper[4956]: I1126 16:56:11.400642 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-89rgc"] Nov 26 16:56:11 crc kubenswrapper[4956]: I1126 16:56:11.410849 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-89rgc" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="registry-server" containerID="cri-o://3ed2b1c5ffe0f3394ca5d11fb4ee10a26bb1708f57f930ac908e9ba00bcce628" gracePeriod=2 Nov 26 16:56:13 crc kubenswrapper[4956]: I1126 16:56:13.243997 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p74bs" event={"ID":"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da","Type":"ContainerStarted","Data":"a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8"} Nov 26 16:56:13 crc kubenswrapper[4956]: I1126 16:56:13.248685 4956 generic.go:334] "Generic (PLEG): container finished" podID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerID="3ed2b1c5ffe0f3394ca5d11fb4ee10a26bb1708f57f930ac908e9ba00bcce628" exitCode=0 Nov 26 16:56:13 crc kubenswrapper[4956]: I1126 16:56:13.248762 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89rgc" event={"ID":"638a6242-d034-4b5d-90f9-f25fcf1bc742","Type":"ContainerDied","Data":"3ed2b1c5ffe0f3394ca5d11fb4ee10a26bb1708f57f930ac908e9ba00bcce628"} Nov 26 16:56:13 crc kubenswrapper[4956]: I1126 16:56:13.252600 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"886bfbf287a7f74a3eeafb773a2ec315c022875497b66a1ecbc8a4ff191c22b3"} Nov 26 16:56:13 crc kubenswrapper[4956]: I1126 16:56:13.264771 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-p74bs" podStartSLOduration=4.680327712 podStartE2EDuration="1m1.264741682s" podCreationTimestamp="2025-11-26 16:55:12 +0000 UTC" firstStartedPulling="2025-11-26 16:55:14.641093633 +0000 UTC m=+160.337054195" lastFinishedPulling="2025-11-26 16:56:11.225507613 +0000 UTC m=+216.921468165" observedRunningTime="2025-11-26 16:56:13.261940791 +0000 UTC m=+218.957901363" watchObservedRunningTime="2025-11-26 16:56:13.264741682 +0000 UTC m=+218.960702234" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.160392 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.265719 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-89rgc" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.265777 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-89rgc" event={"ID":"638a6242-d034-4b5d-90f9-f25fcf1bc742","Type":"ContainerDied","Data":"cc097537545aef922936d70eec2011725e24ff5a3713df41a9fed6410bf4dc16"} Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.265820 4956 scope.go:117] "RemoveContainer" containerID="3ed2b1c5ffe0f3394ca5d11fb4ee10a26bb1708f57f930ac908e9ba00bcce628" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.302243 4956 scope.go:117] "RemoveContainer" containerID="efb0f298147341ab4b0acf25ce1681c0f8a8eadc1f6bf8db610e00bebed5a7c6" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.324463 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6kms5\" (UniqueName: \"kubernetes.io/projected/638a6242-d034-4b5d-90f9-f25fcf1bc742-kube-api-access-6kms5\") pod \"638a6242-d034-4b5d-90f9-f25fcf1bc742\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.324546 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-utilities\") pod \"638a6242-d034-4b5d-90f9-f25fcf1bc742\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.324614 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-catalog-content\") pod \"638a6242-d034-4b5d-90f9-f25fcf1bc742\" (UID: \"638a6242-d034-4b5d-90f9-f25fcf1bc742\") " Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.327542 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-utilities" (OuterVolumeSpecName: "utilities") pod "638a6242-d034-4b5d-90f9-f25fcf1bc742" (UID: "638a6242-d034-4b5d-90f9-f25fcf1bc742"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.330889 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/638a6242-d034-4b5d-90f9-f25fcf1bc742-kube-api-access-6kms5" (OuterVolumeSpecName: "kube-api-access-6kms5") pod "638a6242-d034-4b5d-90f9-f25fcf1bc742" (UID: "638a6242-d034-4b5d-90f9-f25fcf1bc742"). InnerVolumeSpecName "kube-api-access-6kms5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.383287 4956 scope.go:117] "RemoveContainer" containerID="6d1cd7d539adde2b27026f8a3021441a115f9738c68e0e8f36068c3e1ddb8d46" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.423229 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "638a6242-d034-4b5d-90f9-f25fcf1bc742" (UID: "638a6242-d034-4b5d-90f9-f25fcf1bc742"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.426790 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6kms5\" (UniqueName: \"kubernetes.io/projected/638a6242-d034-4b5d-90f9-f25fcf1bc742-kube-api-access-6kms5\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.426829 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.426842 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/638a6242-d034-4b5d-90f9-f25fcf1bc742-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.594056 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-89rgc"] Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.598456 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-89rgc"] Nov 26 16:56:14 crc kubenswrapper[4956]: I1126 16:56:14.944991 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.017728 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" path="/var/lib/kubelet/pods/638a6242-d034-4b5d-90f9-f25fcf1bc742/volumes" Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.018758 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.273193 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ks5k" event={"ID":"3540205e-79fd-4756-8203-d9b06a0ccf12","Type":"ContainerStarted","Data":"f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10"} Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.276302 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerStarted","Data":"219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059"} Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.278661 4956 generic.go:334] "Generic (PLEG): container finished" podID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerID="232a89cbe53418655a89676bf9679b70a6a1fc8209c759f402a9a5e4e8964b8d" exitCode=0 Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.278725 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sj" event={"ID":"9a71e654-5284-4cbb-a3de-59b7e29d1fc9","Type":"ContainerDied","Data":"232a89cbe53418655a89676bf9679b70a6a1fc8209c759f402a9a5e4e8964b8d"} Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.287357 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerStarted","Data":"e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b"} Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.933233 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:56:15 crc kubenswrapper[4956]: I1126 16:56:15.984067 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:56:16 crc kubenswrapper[4956]: I1126 16:56:16.293406 4956 generic.go:334] "Generic (PLEG): container finished" podID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerID="219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059" exitCode=0 Nov 26 16:56:16 crc kubenswrapper[4956]: I1126 16:56:16.293826 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerDied","Data":"219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059"} Nov 26 16:56:16 crc kubenswrapper[4956]: I1126 16:56:16.297580 4956 generic.go:334] "Generic (PLEG): container finished" podID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerID="e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b" exitCode=0 Nov 26 16:56:16 crc kubenswrapper[4956]: I1126 16:56:16.298008 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerDied","Data":"e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b"} Nov 26 16:56:16 crc kubenswrapper[4956]: I1126 16:56:16.360072 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6ks5k" podStartSLOduration=4.727023723 podStartE2EDuration="1m4.360045905s" podCreationTimestamp="2025-11-26 16:55:12 +0000 UTC" firstStartedPulling="2025-11-26 16:55:14.625104933 +0000 UTC m=+160.321065485" lastFinishedPulling="2025-11-26 16:56:14.258127115 +0000 UTC m=+219.954087667" observedRunningTime="2025-11-26 16:56:16.353578029 +0000 UTC m=+222.049538591" watchObservedRunningTime="2025-11-26 16:56:16.360045905 +0000 UTC m=+222.056006457" Nov 26 16:56:17 crc kubenswrapper[4956]: I1126 16:56:17.799235 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2bxq"] Nov 26 16:56:17 crc kubenswrapper[4956]: I1126 16:56:17.801089 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-q2bxq" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="registry-server" containerID="cri-o://381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd" gracePeriod=2 Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.168184 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.283540 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-54jhp\" (UniqueName: \"kubernetes.io/projected/2fafdf53-76b4-49ba-8346-430e2bd5d345-kube-api-access-54jhp\") pod \"2fafdf53-76b4-49ba-8346-430e2bd5d345\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.283630 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-utilities\") pod \"2fafdf53-76b4-49ba-8346-430e2bd5d345\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.283680 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-catalog-content\") pod \"2fafdf53-76b4-49ba-8346-430e2bd5d345\" (UID: \"2fafdf53-76b4-49ba-8346-430e2bd5d345\") " Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.284508 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-utilities" (OuterVolumeSpecName: "utilities") pod "2fafdf53-76b4-49ba-8346-430e2bd5d345" (UID: "2fafdf53-76b4-49ba-8346-430e2bd5d345"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.290791 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fafdf53-76b4-49ba-8346-430e2bd5d345-kube-api-access-54jhp" (OuterVolumeSpecName: "kube-api-access-54jhp") pod "2fafdf53-76b4-49ba-8346-430e2bd5d345" (UID: "2fafdf53-76b4-49ba-8346-430e2bd5d345"). InnerVolumeSpecName "kube-api-access-54jhp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.310076 4956 generic.go:334] "Generic (PLEG): container finished" podID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerID="381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd" exitCode=0 Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.310125 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2bxq" event={"ID":"2fafdf53-76b4-49ba-8346-430e2bd5d345","Type":"ContainerDied","Data":"381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd"} Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.310189 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-q2bxq" event={"ID":"2fafdf53-76b4-49ba-8346-430e2bd5d345","Type":"ContainerDied","Data":"63cf40b760a44d91683e1aa3cdfd56ade9de67a6a5e0898de7554b183fa6ddd9"} Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.310212 4956 scope.go:117] "RemoveContainer" containerID="381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.310218 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-q2bxq" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.317759 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2fafdf53-76b4-49ba-8346-430e2bd5d345" (UID: "2fafdf53-76b4-49ba-8346-430e2bd5d345"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.358409 4956 scope.go:117] "RemoveContainer" containerID="a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.376303 4956 scope.go:117] "RemoveContainer" containerID="9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.385277 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-54jhp\" (UniqueName: \"kubernetes.io/projected/2fafdf53-76b4-49ba-8346-430e2bd5d345-kube-api-access-54jhp\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.385312 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.385327 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2fafdf53-76b4-49ba-8346-430e2bd5d345-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.395071 4956 scope.go:117] "RemoveContainer" containerID="381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd" Nov 26 16:56:18 crc kubenswrapper[4956]: E1126 16:56:18.395666 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd\": container with ID starting with 381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd not found: ID does not exist" containerID="381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.395721 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd"} err="failed to get container status \"381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd\": rpc error: code = NotFound desc = could not find container \"381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd\": container with ID starting with 381eb7ddcb4f379c033d5c6a4d44a7f754c500fc5402c57571dd99ad33d907cd not found: ID does not exist" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.395765 4956 scope.go:117] "RemoveContainer" containerID="a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526" Nov 26 16:56:18 crc kubenswrapper[4956]: E1126 16:56:18.396326 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526\": container with ID starting with a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526 not found: ID does not exist" containerID="a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.396369 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526"} err="failed to get container status \"a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526\": rpc error: code = NotFound desc = could not find container \"a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526\": container with ID starting with a1834173ca78ae3a37b7bf0bfc45ee42e14fcbeaf2170398696d2f3248c0f526 not found: ID does not exist" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.396400 4956 scope.go:117] "RemoveContainer" containerID="9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210" Nov 26 16:56:18 crc kubenswrapper[4956]: E1126 16:56:18.396691 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210\": container with ID starting with 9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210 not found: ID does not exist" containerID="9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.396726 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210"} err="failed to get container status \"9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210\": rpc error: code = NotFound desc = could not find container \"9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210\": container with ID starting with 9f45cd334e20b8b5a29fa31ef7140e76655a9459a5ff58b3bd1ba25b983b1210 not found: ID does not exist" Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.640879 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2bxq"] Nov 26 16:56:18 crc kubenswrapper[4956]: I1126 16:56:18.643198 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-q2bxq"] Nov 26 16:56:19 crc kubenswrapper[4956]: I1126 16:56:19.004971 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" path="/var/lib/kubelet/pods/2fafdf53-76b4-49ba-8346-430e2bd5d345/volumes" Nov 26 16:56:20 crc kubenswrapper[4956]: I1126 16:56:20.340760 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sj" event={"ID":"9a71e654-5284-4cbb-a3de-59b7e29d1fc9","Type":"ContainerStarted","Data":"b4c62e4cc1174d0a23e946a741beb3550a938126f6fb30ce6ce326d7d5af3be4"} Nov 26 16:56:21 crc kubenswrapper[4956]: I1126 16:56:21.372796 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-6z8sj" podStartSLOduration=4.697629907 podStartE2EDuration="1m7.372762644s" podCreationTimestamp="2025-11-26 16:55:14 +0000 UTC" firstStartedPulling="2025-11-26 16:55:15.683375473 +0000 UTC m=+161.379336025" lastFinishedPulling="2025-11-26 16:56:18.35850821 +0000 UTC m=+224.054468762" observedRunningTime="2025-11-26 16:56:21.36668766 +0000 UTC m=+227.062648252" watchObservedRunningTime="2025-11-26 16:56:21.372762644 +0000 UTC m=+227.068723196" Nov 26 16:56:22 crc kubenswrapper[4956]: I1126 16:56:22.531087 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:56:22 crc kubenswrapper[4956]: I1126 16:56:22.531161 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:56:22 crc kubenswrapper[4956]: I1126 16:56:22.593844 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:56:22 crc kubenswrapper[4956]: I1126 16:56:22.952905 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:56:22 crc kubenswrapper[4956]: I1126 16:56:22.953653 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:56:22 crc kubenswrapper[4956]: I1126 16:56:22.993985 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:56:23 crc kubenswrapper[4956]: I1126 16:56:23.400518 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:56:23 crc kubenswrapper[4956]: I1126 16:56:23.403590 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:56:23 crc kubenswrapper[4956]: I1126 16:56:23.800689 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ks5k"] Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.367878 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerStarted","Data":"9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda"} Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.370750 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerStarted","Data":"23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208"} Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.394725 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-cs6v7" podStartSLOduration=3.820358841 podStartE2EDuration="1m12.39469916s" podCreationTimestamp="2025-11-26 16:55:12 +0000 UTC" firstStartedPulling="2025-11-26 16:55:14.601988009 +0000 UTC m=+160.297948561" lastFinishedPulling="2025-11-26 16:56:23.176328318 +0000 UTC m=+228.872288880" observedRunningTime="2025-11-26 16:56:24.391592751 +0000 UTC m=+230.087553303" watchObservedRunningTime="2025-11-26 16:56:24.39469916 +0000 UTC m=+230.090659712" Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.422290 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-hsg7t" podStartSLOduration=4.042992291 podStartE2EDuration="1m12.422265153s" podCreationTimestamp="2025-11-26 16:55:12 +0000 UTC" firstStartedPulling="2025-11-26 16:55:14.621443848 +0000 UTC m=+160.317404400" lastFinishedPulling="2025-11-26 16:56:23.00071669 +0000 UTC m=+228.696677262" observedRunningTime="2025-11-26 16:56:24.419337019 +0000 UTC m=+230.115297571" watchObservedRunningTime="2025-11-26 16:56:24.422265153 +0000 UTC m=+230.118225705" Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.520748 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.521108 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:56:24 crc kubenswrapper[4956]: I1126 16:56:24.567584 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.376975 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6ks5k" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="registry-server" containerID="cri-o://f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10" gracePeriod=2 Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.423360 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.799835 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.805430 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-utilities\") pod \"3540205e-79fd-4756-8203-d9b06a0ccf12\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.805574 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rzmgz\" (UniqueName: \"kubernetes.io/projected/3540205e-79fd-4756-8203-d9b06a0ccf12-kube-api-access-rzmgz\") pod \"3540205e-79fd-4756-8203-d9b06a0ccf12\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.805637 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-catalog-content\") pod \"3540205e-79fd-4756-8203-d9b06a0ccf12\" (UID: \"3540205e-79fd-4756-8203-d9b06a0ccf12\") " Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.806613 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-utilities" (OuterVolumeSpecName: "utilities") pod "3540205e-79fd-4756-8203-d9b06a0ccf12" (UID: "3540205e-79fd-4756-8203-d9b06a0ccf12"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.813387 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3540205e-79fd-4756-8203-d9b06a0ccf12-kube-api-access-rzmgz" (OuterVolumeSpecName: "kube-api-access-rzmgz") pod "3540205e-79fd-4756-8203-d9b06a0ccf12" (UID: "3540205e-79fd-4756-8203-d9b06a0ccf12"). InnerVolumeSpecName "kube-api-access-rzmgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.869331 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "3540205e-79fd-4756-8203-d9b06a0ccf12" (UID: "3540205e-79fd-4756-8203-d9b06a0ccf12"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.906853 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rzmgz\" (UniqueName: \"kubernetes.io/projected/3540205e-79fd-4756-8203-d9b06a0ccf12-kube-api-access-rzmgz\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.906904 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:25 crc kubenswrapper[4956]: I1126 16:56:25.906916 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3540205e-79fd-4756-8203-d9b06a0ccf12-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.384827 4956 generic.go:334] "Generic (PLEG): container finished" podID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerID="f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10" exitCode=0 Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.384895 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ks5k" event={"ID":"3540205e-79fd-4756-8203-d9b06a0ccf12","Type":"ContainerDied","Data":"f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10"} Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.384945 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6ks5k" event={"ID":"3540205e-79fd-4756-8203-d9b06a0ccf12","Type":"ContainerDied","Data":"992fd210af8b1325c3b5f4a501a1d4d8be8746c9ea23d2763eb7891cf57e5df7"} Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.384968 4956 scope.go:117] "RemoveContainer" containerID="f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.384977 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6ks5k" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.403427 4956 scope.go:117] "RemoveContainer" containerID="6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.421191 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6ks5k"] Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.423093 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6ks5k"] Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.424289 4956 scope.go:117] "RemoveContainer" containerID="9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.447768 4956 scope.go:117] "RemoveContainer" containerID="f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10" Nov 26 16:56:26 crc kubenswrapper[4956]: E1126 16:56:26.448595 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10\": container with ID starting with f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10 not found: ID does not exist" containerID="f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.448642 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10"} err="failed to get container status \"f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10\": rpc error: code = NotFound desc = could not find container \"f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10\": container with ID starting with f898ef9ec88f6ad7b96cf5a4d9209701ffc05c1a0eb0b5559f2c9cc5ffaeec10 not found: ID does not exist" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.448683 4956 scope.go:117] "RemoveContainer" containerID="6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2" Nov 26 16:56:26 crc kubenswrapper[4956]: E1126 16:56:26.449104 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2\": container with ID starting with 6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2 not found: ID does not exist" containerID="6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.449133 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2"} err="failed to get container status \"6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2\": rpc error: code = NotFound desc = could not find container \"6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2\": container with ID starting with 6753b6521a1ed5d642a0eded3180d21a6697c39bc99aed842520aa8285af36e2 not found: ID does not exist" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.449152 4956 scope.go:117] "RemoveContainer" containerID="9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40" Nov 26 16:56:26 crc kubenswrapper[4956]: E1126 16:56:26.449458 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40\": container with ID starting with 9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40 not found: ID does not exist" containerID="9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40" Nov 26 16:56:26 crc kubenswrapper[4956]: I1126 16:56:26.449498 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40"} err="failed to get container status \"9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40\": rpc error: code = NotFound desc = could not find container \"9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40\": container with ID starting with 9b937c5b7f39fa0e850d72a48b5b9e8c5ef11df084dca7a0b4f3a56c5a3e2e40 not found: ID does not exist" Nov 26 16:56:27 crc kubenswrapper[4956]: I1126 16:56:27.002310 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" path="/var/lib/kubelet/pods/3540205e-79fd-4756-8203-d9b06a0ccf12/volumes" Nov 26 16:56:30 crc kubenswrapper[4956]: I1126 16:56:30.464353 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerName="oauth-openshift" containerID="cri-o://7d3d6a4e7e389b2af8fdeea87b1322ddd0313dcdfc2f9f43ada86d801a671cd8" gracePeriod=15 Nov 26 16:56:32 crc kubenswrapper[4956]: I1126 16:56:32.424083 4956 generic.go:334] "Generic (PLEG): container finished" podID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerID="7d3d6a4e7e389b2af8fdeea87b1322ddd0313dcdfc2f9f43ada86d801a671cd8" exitCode=0 Nov 26 16:56:32 crc kubenswrapper[4956]: I1126 16:56:32.424316 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" event={"ID":"5552a148-c82e-4afb-9a27-9ce40714bdb0","Type":"ContainerDied","Data":"7d3d6a4e7e389b2af8fdeea87b1322ddd0313dcdfc2f9f43ada86d801a671cd8"} Nov 26 16:56:32 crc kubenswrapper[4956]: I1126 16:56:32.692321 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:56:32 crc kubenswrapper[4956]: I1126 16:56:32.692393 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:56:32 crc kubenswrapper[4956]: I1126 16:56:32.759539 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:56:32 crc kubenswrapper[4956]: I1126 16:56:32.978138 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117248 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-trusted-ca-bundle\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117335 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-service-ca\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117417 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-policies\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117465 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-login\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117503 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-serving-cert\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117545 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-session\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117613 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-cliconfig\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117682 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-dir\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117802 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-idp-0-file-data\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117840 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-error\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117905 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l8lgn\" (UniqueName: \"kubernetes.io/projected/5552a148-c82e-4afb-9a27-9ce40714bdb0-kube-api-access-l8lgn\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117943 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-ocp-branding-template\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.117979 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-provider-selection\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.118045 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-router-certs\") pod \"5552a148-c82e-4afb-9a27-9ce40714bdb0\" (UID: \"5552a148-c82e-4afb-9a27-9ce40714bdb0\") " Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.118099 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.118727 4956 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.119284 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.119347 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.119402 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.120325 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.126363 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.128399 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5552a148-c82e-4afb-9a27-9ce40714bdb0-kube-api-access-l8lgn" (OuterVolumeSpecName: "kube-api-access-l8lgn") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "kube-api-access-l8lgn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.134347 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.134532 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.134844 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.135257 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.135378 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.135809 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.135936 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "5552a148-c82e-4afb-9a27-9ce40714bdb0" (UID: "5552a148-c82e-4afb-9a27-9ce40714bdb0"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.155091 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.155545 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.194458 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221521 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221571 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l8lgn\" (UniqueName: \"kubernetes.io/projected/5552a148-c82e-4afb-9a27-9ce40714bdb0-kube-api-access-l8lgn\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221591 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221615 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221633 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221652 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221667 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221688 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221702 4956 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-audit-policies\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221718 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221735 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221753 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.221767 4956 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/5552a148-c82e-4afb-9a27-9ce40714bdb0-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.432732 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" event={"ID":"5552a148-c82e-4afb-9a27-9ce40714bdb0","Type":"ContainerDied","Data":"59f9a0880b9fdd99828564611d7f38e3ce6e8f06e2baf458618a0a1b55089705"} Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.432819 4956 scope.go:117] "RemoveContainer" containerID="7d3d6a4e7e389b2af8fdeea87b1322ddd0313dcdfc2f9f43ada86d801a671cd8" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.432931 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-8tmqv" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.466294 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8tmqv"] Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.468789 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-8tmqv"] Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.482037 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:56:33 crc kubenswrapper[4956]: I1126 16:56:33.493387 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.062825 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b"] Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064402 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7c3360f5-aa2a-4848-9c3c-7a238f8a761f" containerName="pruner" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064458 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7c3360f5-aa2a-4848-9c3c-7a238f8a761f" containerName="pruner" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064485 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064499 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064520 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="extract-utilities" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064533 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="extract-utilities" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064552 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="extract-content" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064565 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="extract-content" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064585 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064598 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064613 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="da69b41a-6323-4098-830b-e0745a21adc6" containerName="pruner" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064625 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="da69b41a-6323-4098-830b-e0745a21adc6" containerName="pruner" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064646 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="extract-content" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064657 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="extract-content" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064679 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerName="oauth-openshift" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064691 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerName="oauth-openshift" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064718 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064731 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064746 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="extract-content" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064758 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="extract-content" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064773 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="extract-utilities" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064785 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="extract-utilities" Nov 26 16:56:34 crc kubenswrapper[4956]: E1126 16:56:34.064802 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="extract-utilities" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.064813 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="extract-utilities" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065029 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="3540205e-79fd-4756-8203-d9b06a0ccf12" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065055 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="638a6242-d034-4b5d-90f9-f25fcf1bc742" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065073 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="da69b41a-6323-4098-830b-e0745a21adc6" containerName="pruner" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065090 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7c3360f5-aa2a-4848-9c3c-7a238f8a761f" containerName="pruner" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065110 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" containerName="oauth-openshift" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065126 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fafdf53-76b4-49ba-8346-430e2bd5d345" containerName="registry-server" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.065800 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076478 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076537 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076509 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076734 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076811 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076818 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.076823 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.077099 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.077105 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.078260 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.082203 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.084626 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.093795 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b"] Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.096624 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.096862 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.108601 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.239535 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-session\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.239607 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.239805 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240050 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240157 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-audit-dir\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240406 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-audit-policies\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240474 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-login\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240537 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240601 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240708 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240743 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-error\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240853 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240919 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.240961 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xgq29\" (UniqueName: \"kubernetes.io/projected/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-kube-api-access-xgq29\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.342918 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-audit-policies\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343063 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-login\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343114 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343159 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343227 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343264 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-error\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343306 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343345 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343383 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xgq29\" (UniqueName: \"kubernetes.io/projected/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-kube-api-access-xgq29\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343423 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-session\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343460 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343503 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343570 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343629 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-audit-dir\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.343783 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-audit-dir\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.344331 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-audit-policies\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.345151 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-cliconfig\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.346455 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.347642 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-service-ca\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.350300 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.351166 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-session\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.352458 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-error\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.353601 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-template-login\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.353907 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-serving-cert\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.354630 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-router-certs\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.356610 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.356741 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.376260 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xgq29\" (UniqueName: \"kubernetes.io/projected/bff288b5-b067-4ecf-a3a1-c4591a3b6fbb-kube-api-access-xgq29\") pod \"oauth-openshift-6cc8b7fbff-hlj9b\" (UID: \"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb\") " pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.390650 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.605990 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cs6v7"] Nov 26 16:56:34 crc kubenswrapper[4956]: I1126 16:56:34.886139 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b"] Nov 26 16:56:35 crc kubenswrapper[4956]: I1126 16:56:35.006021 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5552a148-c82e-4afb-9a27-9ce40714bdb0" path="/var/lib/kubelet/pods/5552a148-c82e-4afb-9a27-9ce40714bdb0/volumes" Nov 26 16:56:35 crc kubenswrapper[4956]: I1126 16:56:35.454859 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" event={"ID":"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb","Type":"ContainerStarted","Data":"69a5caae9ec9688500121bb48413b670dd815067f6d01c2c1bb2181014b9eb96"} Nov 26 16:56:36 crc kubenswrapper[4956]: I1126 16:56:36.464396 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" event={"ID":"bff288b5-b067-4ecf-a3a1-c4591a3b6fbb","Type":"ContainerStarted","Data":"493923d0fbab5cad08b30001970d852900947ed1937b1dfd4333316941d85dd5"} Nov 26 16:56:36 crc kubenswrapper[4956]: I1126 16:56:36.464697 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-cs6v7" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="registry-server" containerID="cri-o://9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda" gracePeriod=2 Nov 26 16:56:36 crc kubenswrapper[4956]: I1126 16:56:36.510925 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" podStartSLOduration=31.510891198 podStartE2EDuration="31.510891198s" podCreationTimestamp="2025-11-26 16:56:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:56:36.506242694 +0000 UTC m=+242.202203286" watchObservedRunningTime="2025-11-26 16:56:36.510891198 +0000 UTC m=+242.206851790" Nov 26 16:56:36 crc kubenswrapper[4956]: I1126 16:56:36.946256 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.087624 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-utilities\") pod \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.087677 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lkcgt\" (UniqueName: \"kubernetes.io/projected/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-kube-api-access-lkcgt\") pod \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.087707 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-catalog-content\") pod \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\" (UID: \"f1fdb8dd-363c-42b9-ab6e-862d8b459f90\") " Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.093070 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-utilities" (OuterVolumeSpecName: "utilities") pod "f1fdb8dd-363c-42b9-ab6e-862d8b459f90" (UID: "f1fdb8dd-363c-42b9-ab6e-862d8b459f90"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.097940 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-kube-api-access-lkcgt" (OuterVolumeSpecName: "kube-api-access-lkcgt") pod "f1fdb8dd-363c-42b9-ab6e-862d8b459f90" (UID: "f1fdb8dd-363c-42b9-ab6e-862d8b459f90"). InnerVolumeSpecName "kube-api-access-lkcgt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.147230 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f1fdb8dd-363c-42b9-ab6e-862d8b459f90" (UID: "f1fdb8dd-363c-42b9-ab6e-862d8b459f90"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.189669 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.190742 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.190810 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lkcgt\" (UniqueName: \"kubernetes.io/projected/f1fdb8dd-363c-42b9-ab6e-862d8b459f90-kube-api-access-lkcgt\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.478770 4956 generic.go:334] "Generic (PLEG): container finished" podID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerID="9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda" exitCode=0 Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.478892 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-cs6v7" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.478926 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerDied","Data":"9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda"} Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.479030 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-cs6v7" event={"ID":"f1fdb8dd-363c-42b9-ab6e-862d8b459f90","Type":"ContainerDied","Data":"2361cfcdc662d59e2d967b8f17800be51dd17ec669bb33314a0ba8c48661ed89"} Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.479074 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.479079 4956 scope.go:117] "RemoveContainer" containerID="9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.487607 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-6cc8b7fbff-hlj9b" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.530455 4956 scope.go:117] "RemoveContainer" containerID="219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.552665 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-cs6v7"] Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.579169 4956 scope.go:117] "RemoveContainer" containerID="9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.581626 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-cs6v7"] Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.613109 4956 scope.go:117] "RemoveContainer" containerID="9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda" Nov 26 16:56:37 crc kubenswrapper[4956]: E1126 16:56:37.617023 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda\": container with ID starting with 9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda not found: ID does not exist" containerID="9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.617072 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda"} err="failed to get container status \"9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda\": rpc error: code = NotFound desc = could not find container \"9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda\": container with ID starting with 9e8e63f95aa39870e6c08bb312c5f19b84cf3bcdd5edfc807afa1b16d8223dda not found: ID does not exist" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.617101 4956 scope.go:117] "RemoveContainer" containerID="219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059" Nov 26 16:56:37 crc kubenswrapper[4956]: E1126 16:56:37.618304 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059\": container with ID starting with 219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059 not found: ID does not exist" containerID="219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.618331 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059"} err="failed to get container status \"219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059\": rpc error: code = NotFound desc = could not find container \"219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059\": container with ID starting with 219368be5c50f1b9820f02ec9d226ea7d3ef1632dd959265ba529c2c0771e059 not found: ID does not exist" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.618346 4956 scope.go:117] "RemoveContainer" containerID="9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b" Nov 26 16:56:37 crc kubenswrapper[4956]: E1126 16:56:37.618650 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b\": container with ID starting with 9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b not found: ID does not exist" containerID="9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b" Nov 26 16:56:37 crc kubenswrapper[4956]: I1126 16:56:37.618670 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b"} err="failed to get container status \"9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b\": rpc error: code = NotFound desc = could not find container \"9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b\": container with ID starting with 9511913bbcc53ac3c0fd6dd6b7f8c1417b1bc5835094954afb8e93edd047e72b not found: ID does not exist" Nov 26 16:56:39 crc kubenswrapper[4956]: I1126 16:56:39.007638 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" path="/var/lib/kubelet/pods/f1fdb8dd-363c-42b9-ab6e-862d8b459f90/volumes" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.460718 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p74bs"] Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.462099 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-p74bs" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="registry-server" containerID="cri-o://a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8" gracePeriod=30 Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.475540 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hsg7t"] Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.476059 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-hsg7t" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="registry-server" containerID="cri-o://23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208" gracePeriod=30 Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.485930 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwzxf"] Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.486407 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" containerID="cri-o://f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0" gracePeriod=30 Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.505354 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sj"] Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.505788 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-6z8sj" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="registry-server" containerID="cri-o://b4c62e4cc1174d0a23e946a741beb3550a938126f6fb30ce6ce326d7d5af3be4" gracePeriod=30 Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.517209 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8drrp"] Nov 26 16:56:56 crc kubenswrapper[4956]: E1126 16:56:56.517695 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="extract-utilities" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.517745 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="extract-utilities" Nov 26 16:56:56 crc kubenswrapper[4956]: E1126 16:56:56.517773 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="registry-server" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.517799 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="registry-server" Nov 26 16:56:56 crc kubenswrapper[4956]: E1126 16:56:56.517819 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="extract-content" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.517828 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="extract-content" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.518011 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f1fdb8dd-363c-42b9-ab6e-862d8b459f90" containerName="registry-server" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.518883 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.520000 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zc7rm"] Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.520329 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zc7rm" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="registry-server" containerID="cri-o://a5a3ef711b91179e96ea3b6f2bf744d56766cb21a99cf44f989b96e1dff6587c" gracePeriod=30 Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.525725 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8drrp"] Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.628929 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c33f19c2-c49b-4677-805f-e60441939c00-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.628983 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c33f19c2-c49b-4677-805f-e60441939c00-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.629174 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dwz9q\" (UniqueName: \"kubernetes.io/projected/c33f19c2-c49b-4677-805f-e60441939c00-kube-api-access-dwz9q\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.730258 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c33f19c2-c49b-4677-805f-e60441939c00-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.730768 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c33f19c2-c49b-4677-805f-e60441939c00-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.731160 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dwz9q\" (UniqueName: \"kubernetes.io/projected/c33f19c2-c49b-4677-805f-e60441939c00-kube-api-access-dwz9q\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.731793 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/c33f19c2-c49b-4677-805f-e60441939c00-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.738288 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/c33f19c2-c49b-4677-805f-e60441939c00-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.746988 4956 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-bwzxf container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" start-of-body= Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.747064 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.38:8080/healthz\": dial tcp 10.217.0.38:8080: connect: connection refused" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.748724 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dwz9q\" (UniqueName: \"kubernetes.io/projected/c33f19c2-c49b-4677-805f-e60441939c00-kube-api-access-dwz9q\") pod \"marketplace-operator-79b997595-8drrp\" (UID: \"c33f19c2-c49b-4677-805f-e60441939c00\") " pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:56 crc kubenswrapper[4956]: I1126 16:56:56.950181 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.176518 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-8drrp"] Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.443309 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.556996 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.570687 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.625438 4956 generic.go:334] "Generic (PLEG): container finished" podID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerID="a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8" exitCode=0 Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.625540 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p74bs" event={"ID":"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da","Type":"ContainerDied","Data":"a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.625581 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-p74bs" event={"ID":"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da","Type":"ContainerDied","Data":"bba620648aed64b10b78eee8a3d0e7f4dc00d8988efb1802b3bffcc8f04cbdff"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.625607 4956 scope.go:117] "RemoveContainer" containerID="a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.625803 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-p74bs" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.631624 4956 generic.go:334] "Generic (PLEG): container finished" podID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerID="b4c62e4cc1174d0a23e946a741beb3550a938126f6fb30ce6ce326d7d5af3be4" exitCode=0 Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.631724 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sj" event={"ID":"9a71e654-5284-4cbb-a3de-59b7e29d1fc9","Type":"ContainerDied","Data":"b4c62e4cc1174d0a23e946a741beb3550a938126f6fb30ce6ce326d7d5af3be4"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.639610 4956 generic.go:334] "Generic (PLEG): container finished" podID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerID="a5a3ef711b91179e96ea3b6f2bf744d56766cb21a99cf44f989b96e1dff6587c" exitCode=0 Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.639696 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerDied","Data":"a5a3ef711b91179e96ea3b6f2bf744d56766cb21a99cf44f989b96e1dff6587c"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.642587 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-trusted-ca\") pod \"dead528a-47b6-44fc-97bf-17c308cc2033\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.642674 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hhcq2\" (UniqueName: \"kubernetes.io/projected/dead528a-47b6-44fc-97bf-17c308cc2033-kube-api-access-hhcq2\") pod \"dead528a-47b6-44fc-97bf-17c308cc2033\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.642755 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-operator-metrics\") pod \"dead528a-47b6-44fc-97bf-17c308cc2033\" (UID: \"dead528a-47b6-44fc-97bf-17c308cc2033\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.643118 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" event={"ID":"c33f19c2-c49b-4677-805f-e60441939c00","Type":"ContainerStarted","Data":"56a5dc3481cc85ae171d657dab510dacbb51864c65f262bf710749909d6d1b4c"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.643163 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" event={"ID":"c33f19c2-c49b-4677-805f-e60441939c00","Type":"ContainerStarted","Data":"e41c83e18a58dc64e47bbae75a4c3d83cbd872ccaebc7edb7babd0193e6403ff"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.643430 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-catalog-content\") pod \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.643446 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.644588 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "dead528a-47b6-44fc-97bf-17c308cc2033" (UID: "dead528a-47b6-44fc-97bf-17c308cc2033"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.652241 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "dead528a-47b6-44fc-97bf-17c308cc2033" (UID: "dead528a-47b6-44fc-97bf-17c308cc2033"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.652359 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dead528a-47b6-44fc-97bf-17c308cc2033-kube-api-access-hhcq2" (OuterVolumeSpecName: "kube-api-access-hhcq2") pod "dead528a-47b6-44fc-97bf-17c308cc2033" (UID: "dead528a-47b6-44fc-97bf-17c308cc2033"). InnerVolumeSpecName "kube-api-access-hhcq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.652510 4956 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-8drrp container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" start-of-body= Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.652675 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" podUID="c33f19c2-c49b-4677-805f-e60441939c00" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.56:8080/healthz\": dial tcp 10.217.0.56:8080: connect: connection refused" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.656430 4956 generic.go:334] "Generic (PLEG): container finished" podID="dead528a-47b6-44fc-97bf-17c308cc2033" containerID="f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0" exitCode=0 Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.656509 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" event={"ID":"dead528a-47b6-44fc-97bf-17c308cc2033","Type":"ContainerDied","Data":"f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.656540 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" event={"ID":"dead528a-47b6-44fc-97bf-17c308cc2033","Type":"ContainerDied","Data":"16eb9510dfe9505048b0ed0c5b9f2fedb403dc8428a54678ac215f66f4342223"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.656598 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-bwzxf" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.664266 4956 scope.go:117] "RemoveContainer" containerID="1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.673266 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" podStartSLOduration=1.673239309 podStartE2EDuration="1.673239309s" podCreationTimestamp="2025-11-26 16:56:56 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 16:56:57.666229368 +0000 UTC m=+263.362189920" watchObservedRunningTime="2025-11-26 16:56:57.673239309 +0000 UTC m=+263.369199861" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.678049 4956 generic.go:334] "Generic (PLEG): container finished" podID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerID="23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208" exitCode=0 Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.678092 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerDied","Data":"23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.678118 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-hsg7t" event={"ID":"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732","Type":"ContainerDied","Data":"6dc1d3bdf8c56e1b6c0f73d9dd78d456328faf723b6dca66fc1a5a3f6af8eec1"} Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.678186 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-hsg7t" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.733234 4956 scope.go:117] "RemoveContainer" containerID="e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.736190 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" (UID: "a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746276 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wckpr\" (UniqueName: \"kubernetes.io/projected/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-kube-api-access-wckpr\") pod \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746333 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-catalog-content\") pod \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746360 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-utilities\") pod \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746403 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-utilities\") pod \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\" (UID: \"a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746458 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gsv82\" (UniqueName: \"kubernetes.io/projected/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-kube-api-access-gsv82\") pod \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\" (UID: \"5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746744 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hhcq2\" (UniqueName: \"kubernetes.io/projected/dead528a-47b6-44fc-97bf-17c308cc2033-kube-api-access-hhcq2\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746762 4956 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746774 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.746788 4956 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/dead528a-47b6-44fc-97bf-17c308cc2033-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.748041 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwzxf"] Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.751079 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-bwzxf"] Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.752077 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.753426 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-utilities" (OuterVolumeSpecName: "utilities") pod "5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" (UID: "5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.753531 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-utilities" (OuterVolumeSpecName: "utilities") pod "a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" (UID: "a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.753658 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.756286 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-kube-api-access-gsv82" (OuterVolumeSpecName: "kube-api-access-gsv82") pod "5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" (UID: "5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da"). InnerVolumeSpecName "kube-api-access-gsv82". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.757496 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-kube-api-access-wckpr" (OuterVolumeSpecName: "kube-api-access-wckpr") pod "a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" (UID: "a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732"). InnerVolumeSpecName "kube-api-access-wckpr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.768552 4956 scope.go:117] "RemoveContainer" containerID="a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.774262 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8\": container with ID starting with a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8 not found: ID does not exist" containerID="a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.774328 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8"} err="failed to get container status \"a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8\": rpc error: code = NotFound desc = could not find container \"a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8\": container with ID starting with a93c95df87a125e86f83059eebadd16f7e51b2f97422049144aa7927bbbf65b8 not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.774371 4956 scope.go:117] "RemoveContainer" containerID="1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.775129 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec\": container with ID starting with 1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec not found: ID does not exist" containerID="1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.775158 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec"} err="failed to get container status \"1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec\": rpc error: code = NotFound desc = could not find container \"1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec\": container with ID starting with 1ce1322cc0bbae16f483532a7a056e47562670d393521cbe2daf376e0da648ec not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.775175 4956 scope.go:117] "RemoveContainer" containerID="e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.778102 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245\": container with ID starting with e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245 not found: ID does not exist" containerID="e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.778161 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245"} err="failed to get container status \"e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245\": rpc error: code = NotFound desc = could not find container \"e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245\": container with ID starting with e4e31fcb2ffdb857ce29f3545088da6f5e4541eb07e9d1dceb371771a52d1245 not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.778198 4956 scope.go:117] "RemoveContainer" containerID="f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.813097 4956 scope.go:117] "RemoveContainer" containerID="f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.814780 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0\": container with ID starting with f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0 not found: ID does not exist" containerID="f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.814828 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0"} err="failed to get container status \"f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0\": rpc error: code = NotFound desc = could not find container \"f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0\": container with ID starting with f6dd92f2b2bc04dfa5e16973a4c5733672b8ca90f3102d8d9527d90ac8f2bfd0 not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.814880 4956 scope.go:117] "RemoveContainer" containerID="23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.819791 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" (UID: "5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.845408 4956 scope.go:117] "RemoveContainer" containerID="e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.847759 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.847786 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gsv82\" (UniqueName: \"kubernetes.io/projected/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-kube-api-access-gsv82\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.847798 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wckpr\" (UniqueName: \"kubernetes.io/projected/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732-kube-api-access-wckpr\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.847821 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.847830 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.889387 4956 scope.go:117] "RemoveContainer" containerID="dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.914271 4956 scope.go:117] "RemoveContainer" containerID="23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.914973 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208\": container with ID starting with 23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208 not found: ID does not exist" containerID="23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.915027 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208"} err="failed to get container status \"23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208\": rpc error: code = NotFound desc = could not find container \"23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208\": container with ID starting with 23b7132363a40adda15011b8797bc30437cf31c1ea5d6eddf61669dfccd42208 not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.915063 4956 scope.go:117] "RemoveContainer" containerID="e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.915638 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b\": container with ID starting with e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b not found: ID does not exist" containerID="e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.915679 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b"} err="failed to get container status \"e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b\": rpc error: code = NotFound desc = could not find container \"e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b\": container with ID starting with e48a2b2bd78784eef964e627414caddc5acde5ef1392279814e8090da39a2f3b not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.915714 4956 scope.go:117] "RemoveContainer" containerID="dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8" Nov 26 16:56:57 crc kubenswrapper[4956]: E1126 16:56:57.916086 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8\": container with ID starting with dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8 not found: ID does not exist" containerID="dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.916115 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8"} err="failed to get container status \"dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8\": rpc error: code = NotFound desc = could not find container \"dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8\": container with ID starting with dcef71953e19dc1e700fe6cf3c030c863286d7d50d4bf9990441929bd25b9df8 not found: ID does not exist" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.949101 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-catalog-content\") pod \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.949185 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-utilities\") pod \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.949244 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-catalog-content\") pod \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.949296 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-59dhh\" (UniqueName: \"kubernetes.io/projected/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-kube-api-access-59dhh\") pod \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\" (UID: \"9a71e654-5284-4cbb-a3de-59b7e29d1fc9\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.949348 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-utilities\") pod \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.949398 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6jd9\" (UniqueName: \"kubernetes.io/projected/8aa59c7b-5919-4987-b0d4-cc4a629ee006-kube-api-access-g6jd9\") pod \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\" (UID: \"8aa59c7b-5919-4987-b0d4-cc4a629ee006\") " Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.950225 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-utilities" (OuterVolumeSpecName: "utilities") pod "9a71e654-5284-4cbb-a3de-59b7e29d1fc9" (UID: "9a71e654-5284-4cbb-a3de-59b7e29d1fc9"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.951536 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-utilities" (OuterVolumeSpecName: "utilities") pod "8aa59c7b-5919-4987-b0d4-cc4a629ee006" (UID: "8aa59c7b-5919-4987-b0d4-cc4a629ee006"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.953214 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8aa59c7b-5919-4987-b0d4-cc4a629ee006-kube-api-access-g6jd9" (OuterVolumeSpecName: "kube-api-access-g6jd9") pod "8aa59c7b-5919-4987-b0d4-cc4a629ee006" (UID: "8aa59c7b-5919-4987-b0d4-cc4a629ee006"). InnerVolumeSpecName "kube-api-access-g6jd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.966043 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-kube-api-access-59dhh" (OuterVolumeSpecName: "kube-api-access-59dhh") pod "9a71e654-5284-4cbb-a3de-59b7e29d1fc9" (UID: "9a71e654-5284-4cbb-a3de-59b7e29d1fc9"). InnerVolumeSpecName "kube-api-access-59dhh". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.970995 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-p74bs"] Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.971835 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-p74bs"] Nov 26 16:56:57 crc kubenswrapper[4956]: I1126 16:56:57.979393 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "9a71e654-5284-4cbb-a3de-59b7e29d1fc9" (UID: "9a71e654-5284-4cbb-a3de-59b7e29d1fc9"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.007342 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-hsg7t"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.031081 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-hsg7t"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.050657 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-59dhh\" (UniqueName: \"kubernetes.io/projected/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-kube-api-access-59dhh\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.050698 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.050711 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6jd9\" (UniqueName: \"kubernetes.io/projected/8aa59c7b-5919-4987-b0d4-cc4a629ee006-kube-api-access-g6jd9\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.050721 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.050731 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/9a71e654-5284-4cbb-a3de-59b7e29d1fc9-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.074617 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8aa59c7b-5919-4987-b0d4-cc4a629ee006" (UID: "8aa59c7b-5919-4987-b0d4-cc4a629ee006"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.152495 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8aa59c7b-5919-4987-b0d4-cc4a629ee006-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.686489 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-6z8sj" event={"ID":"9a71e654-5284-4cbb-a3de-59b7e29d1fc9","Type":"ContainerDied","Data":"8c4e6e54fadd1c1e5f40650a7445eff3fea2898524334d102e83b297e11334a3"} Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.686573 4956 scope.go:117] "RemoveContainer" containerID="b4c62e4cc1174d0a23e946a741beb3550a938126f6fb30ce6ce326d7d5af3be4" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.686585 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-6z8sj" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.691244 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zc7rm" event={"ID":"8aa59c7b-5919-4987-b0d4-cc4a629ee006","Type":"ContainerDied","Data":"3747b156728be566238564e674ba49104ae4337d57b13baedf4742ed09afac42"} Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.691334 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zc7rm" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.697700 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-8drrp" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.734308 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zc7rm"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.739675 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zc7rm"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.792913 4956 scope.go:117] "RemoveContainer" containerID="232a89cbe53418655a89676bf9679b70a6a1fc8209c759f402a9a5e4e8964b8d" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.809210 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sj"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.816222 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-6z8sj"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.846381 4956 scope.go:117] "RemoveContainer" containerID="241647a8d8ad17f8dbd7f8adbcc13858bf96e43e3a7ac174a29bbd741500d3ad" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.860128 4956 scope.go:117] "RemoveContainer" containerID="a5a3ef711b91179e96ea3b6f2bf744d56766cb21a99cf44f989b96e1dff6587c" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877319 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-kspnz"] Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877539 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877553 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877564 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877570 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877577 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877584 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877594 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877600 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877613 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877620 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877629 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877636 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877644 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877652 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877660 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877666 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="extract-utilities" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877674 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877681 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877687 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877693 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877704 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877711 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="extract-content" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877717 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877724 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: E1126 16:56:58.877731 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877737 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877847 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877882 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877891 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" containerName="marketplace-operator" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877906 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.877913 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" containerName="registry-server" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.878710 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.881795 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.886927 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kspnz"] Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.888050 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520be241-f85c-43e0-8fe0-892e0df6c66c-catalog-content\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.888211 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-22dbd\" (UniqueName: \"kubernetes.io/projected/520be241-f85c-43e0-8fe0-892e0df6c66c-kube-api-access-22dbd\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.888335 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520be241-f85c-43e0-8fe0-892e0df6c66c-utilities\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.890116 4956 scope.go:117] "RemoveContainer" containerID="c843b0880e491a180db8b43d7cfc56ccfa6cf86c49b4bc95b98271546ffa1021" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.909405 4956 scope.go:117] "RemoveContainer" containerID="1fe2cfe8cfbfd5ce0e736a894a46405247b47b744a903aaa198283a2d1ade092" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.989560 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520be241-f85c-43e0-8fe0-892e0df6c66c-utilities\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.989742 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520be241-f85c-43e0-8fe0-892e0df6c66c-catalog-content\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.989784 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-22dbd\" (UniqueName: \"kubernetes.io/projected/520be241-f85c-43e0-8fe0-892e0df6c66c-kube-api-access-22dbd\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.990220 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/520be241-f85c-43e0-8fe0-892e0df6c66c-utilities\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:58 crc kubenswrapper[4956]: I1126 16:56:58.990351 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/520be241-f85c-43e0-8fe0-892e0df6c66c-catalog-content\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.002361 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da" path="/var/lib/kubelet/pods/5c8dbec8-3c06-48e5-bc95-f4b6b1d2f1da/volumes" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.003117 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8aa59c7b-5919-4987-b0d4-cc4a629ee006" path="/var/lib/kubelet/pods/8aa59c7b-5919-4987-b0d4-cc4a629ee006/volumes" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.003748 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9a71e654-5284-4cbb-a3de-59b7e29d1fc9" path="/var/lib/kubelet/pods/9a71e654-5284-4cbb-a3de-59b7e29d1fc9/volumes" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.004899 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732" path="/var/lib/kubelet/pods/a586fcfa-6e5e-40ff-b8b7-bc2f26f0b732/volumes" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.005650 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dead528a-47b6-44fc-97bf-17c308cc2033" path="/var/lib/kubelet/pods/dead528a-47b6-44fc-97bf-17c308cc2033/volumes" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.006097 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-22dbd\" (UniqueName: \"kubernetes.io/projected/520be241-f85c-43e0-8fe0-892e0df6c66c-kube-api-access-22dbd\") pod \"certified-operators-kspnz\" (UID: \"520be241-f85c-43e0-8fe0-892e0df6c66c\") " pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.205960 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.430440 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-kspnz"] Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.701910 4956 generic.go:334] "Generic (PLEG): container finished" podID="520be241-f85c-43e0-8fe0-892e0df6c66c" containerID="1e1c7801ceb28a0773bd9c80de57b38fcbb9daa06820ec280ce771163c42a660" exitCode=0 Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.702024 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kspnz" event={"ID":"520be241-f85c-43e0-8fe0-892e0df6c66c","Type":"ContainerDied","Data":"1e1c7801ceb28a0773bd9c80de57b38fcbb9daa06820ec280ce771163c42a660"} Nov 26 16:56:59 crc kubenswrapper[4956]: I1126 16:56:59.702095 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kspnz" event={"ID":"520be241-f85c-43e0-8fe0-892e0df6c66c","Type":"ContainerStarted","Data":"1214f3dafde3eedd35dedb220e66f250456a06db0b09f113a4eaa88ed64d05bc"} Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.681997 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mvl9d"] Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.687283 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.690378 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.695836 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mvl9d"] Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.713799 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab83ee79-8dfe-4cd9-b104-49ac79683da9-utilities\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.713906 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab83ee79-8dfe-4cd9-b104-49ac79683da9-catalog-content\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.713968 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-986sr\" (UniqueName: \"kubernetes.io/projected/ab83ee79-8dfe-4cd9-b104-49ac79683da9-kube-api-access-986sr\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.814824 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab83ee79-8dfe-4cd9-b104-49ac79683da9-utilities\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.814910 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab83ee79-8dfe-4cd9-b104-49ac79683da9-catalog-content\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.814962 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-986sr\" (UniqueName: \"kubernetes.io/projected/ab83ee79-8dfe-4cd9-b104-49ac79683da9-kube-api-access-986sr\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.815507 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/ab83ee79-8dfe-4cd9-b104-49ac79683da9-utilities\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.816084 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/ab83ee79-8dfe-4cd9-b104-49ac79683da9-catalog-content\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:00 crc kubenswrapper[4956]: I1126 16:57:00.836545 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-986sr\" (UniqueName: \"kubernetes.io/projected/ab83ee79-8dfe-4cd9-b104-49ac79683da9-kube-api-access-986sr\") pod \"redhat-operators-mvl9d\" (UID: \"ab83ee79-8dfe-4cd9-b104-49ac79683da9\") " pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.009106 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.279261 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dhxg8"] Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.280966 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.285312 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.289243 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhxg8"] Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.428343 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9pbsn\" (UniqueName: \"kubernetes.io/projected/f5bfa256-d54e-4aa1-a63c-7805fb956688-kube-api-access-9pbsn\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.428426 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-catalog-content\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.428490 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-utilities\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.437477 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mvl9d"] Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.529538 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-utilities\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.529606 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9pbsn\" (UniqueName: \"kubernetes.io/projected/f5bfa256-d54e-4aa1-a63c-7805fb956688-kube-api-access-9pbsn\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.529666 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-catalog-content\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.530413 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-catalog-content\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.530699 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-utilities\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.550314 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9pbsn\" (UniqueName: \"kubernetes.io/projected/f5bfa256-d54e-4aa1-a63c-7805fb956688-kube-api-access-9pbsn\") pod \"community-operators-dhxg8\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.606114 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.724530 4956 generic.go:334] "Generic (PLEG): container finished" podID="ab83ee79-8dfe-4cd9-b104-49ac79683da9" containerID="0a7f0c23d22e9d52a34044de2def76111e49c3367e57e17cc2d8ec6abfcc05ff" exitCode=0 Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.724585 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvl9d" event={"ID":"ab83ee79-8dfe-4cd9-b104-49ac79683da9","Type":"ContainerDied","Data":"0a7f0c23d22e9d52a34044de2def76111e49c3367e57e17cc2d8ec6abfcc05ff"} Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.724616 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvl9d" event={"ID":"ab83ee79-8dfe-4cd9-b104-49ac79683da9","Type":"ContainerStarted","Data":"2bc43aba6287b38b9b6f2e82e0a0e7774c9ded9d42d38ad4c34cfb359af5c982"} Nov 26 16:57:01 crc kubenswrapper[4956]: I1126 16:57:01.892935 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dhxg8"] Nov 26 16:57:02 crc kubenswrapper[4956]: I1126 16:57:02.732468 4956 generic.go:334] "Generic (PLEG): container finished" podID="520be241-f85c-43e0-8fe0-892e0df6c66c" containerID="54eb21d5a2d4897190ddacb1cc9d9f2ebb1c5217e2fd2b5ec22b927c1f5ce307" exitCode=0 Nov 26 16:57:02 crc kubenswrapper[4956]: I1126 16:57:02.732533 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kspnz" event={"ID":"520be241-f85c-43e0-8fe0-892e0df6c66c","Type":"ContainerDied","Data":"54eb21d5a2d4897190ddacb1cc9d9f2ebb1c5217e2fd2b5ec22b927c1f5ce307"} Nov 26 16:57:02 crc kubenswrapper[4956]: I1126 16:57:02.734727 4956 generic.go:334] "Generic (PLEG): container finished" podID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerID="2e3fee9b24fa6d80c28e4b559c97c679a25644f01dc8aa71dc43aab2ef50c19f" exitCode=0 Nov 26 16:57:02 crc kubenswrapper[4956]: I1126 16:57:02.734758 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerDied","Data":"2e3fee9b24fa6d80c28e4b559c97c679a25644f01dc8aa71dc43aab2ef50c19f"} Nov 26 16:57:02 crc kubenswrapper[4956]: I1126 16:57:02.734778 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerStarted","Data":"9674a2da22440107edab1c13ca84e9e65457fa9fa4020b74c937273296a4b37d"} Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.082005 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-c6bhv"] Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.085967 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.089636 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.097092 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6bhv"] Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.255623 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abad77b8-1ee5-455d-9510-f76fbc9f3014-utilities\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.256100 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-797lj\" (UniqueName: \"kubernetes.io/projected/abad77b8-1ee5-455d-9510-f76fbc9f3014-kube-api-access-797lj\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.256148 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abad77b8-1ee5-455d-9510-f76fbc9f3014-catalog-content\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.361117 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abad77b8-1ee5-455d-9510-f76fbc9f3014-utilities\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.361188 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-797lj\" (UniqueName: \"kubernetes.io/projected/abad77b8-1ee5-455d-9510-f76fbc9f3014-kube-api-access-797lj\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.361230 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abad77b8-1ee5-455d-9510-f76fbc9f3014-catalog-content\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.361632 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/abad77b8-1ee5-455d-9510-f76fbc9f3014-utilities\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.361728 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/abad77b8-1ee5-455d-9510-f76fbc9f3014-catalog-content\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.386764 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-797lj\" (UniqueName: \"kubernetes.io/projected/abad77b8-1ee5-455d-9510-f76fbc9f3014-kube-api-access-797lj\") pod \"redhat-marketplace-c6bhv\" (UID: \"abad77b8-1ee5-455d-9510-f76fbc9f3014\") " pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.414352 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.753721 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvl9d" event={"ID":"ab83ee79-8dfe-4cd9-b104-49ac79683da9","Type":"ContainerStarted","Data":"fb8acc2487168bf4787e0223f9b95c7d4bfc7f1751d42f3ce13e8748fd55ab9b"} Nov 26 16:57:03 crc kubenswrapper[4956]: I1126 16:57:03.874147 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-c6bhv"] Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.758206 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-kspnz" event={"ID":"520be241-f85c-43e0-8fe0-892e0df6c66c","Type":"ContainerStarted","Data":"c16dce992bd2a860b0e1b968fb7478ae5d925f1ef2ffcad89a025607c1f1e911"} Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.760939 4956 generic.go:334] "Generic (PLEG): container finished" podID="ab83ee79-8dfe-4cd9-b104-49ac79683da9" containerID="fb8acc2487168bf4787e0223f9b95c7d4bfc7f1751d42f3ce13e8748fd55ab9b" exitCode=0 Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.761007 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvl9d" event={"ID":"ab83ee79-8dfe-4cd9-b104-49ac79683da9","Type":"ContainerDied","Data":"fb8acc2487168bf4787e0223f9b95c7d4bfc7f1751d42f3ce13e8748fd55ab9b"} Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.764917 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerStarted","Data":"551d104c95a65107374d63743e3464cf6c93f4d7b1a067474d38a1de5641d4bd"} Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.766323 4956 generic.go:334] "Generic (PLEG): container finished" podID="abad77b8-1ee5-455d-9510-f76fbc9f3014" containerID="8f8d3ac50aaba926d153047eb1b3fd3531a673046eba427ae068abe5528e6283" exitCode=0 Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.766363 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6bhv" event={"ID":"abad77b8-1ee5-455d-9510-f76fbc9f3014","Type":"ContainerDied","Data":"8f8d3ac50aaba926d153047eb1b3fd3531a673046eba427ae068abe5528e6283"} Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.766384 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6bhv" event={"ID":"abad77b8-1ee5-455d-9510-f76fbc9f3014","Type":"ContainerStarted","Data":"ec0de68a065379f453e96c01e942239988c3c91cc97eee635804e0902867703c"} Nov 26 16:57:04 crc kubenswrapper[4956]: I1126 16:57:04.781798 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-kspnz" podStartSLOduration=2.560895039 podStartE2EDuration="6.781777197s" podCreationTimestamp="2025-11-26 16:56:58 +0000 UTC" firstStartedPulling="2025-11-26 16:56:59.705365736 +0000 UTC m=+265.401326288" lastFinishedPulling="2025-11-26 16:57:03.926247894 +0000 UTC m=+269.622208446" observedRunningTime="2025-11-26 16:57:04.776463465 +0000 UTC m=+270.472424027" watchObservedRunningTime="2025-11-26 16:57:04.781777197 +0000 UTC m=+270.477737749" Nov 26 16:57:05 crc kubenswrapper[4956]: I1126 16:57:05.775745 4956 generic.go:334] "Generic (PLEG): container finished" podID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerID="551d104c95a65107374d63743e3464cf6c93f4d7b1a067474d38a1de5641d4bd" exitCode=0 Nov 26 16:57:05 crc kubenswrapper[4956]: I1126 16:57:05.775893 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerDied","Data":"551d104c95a65107374d63743e3464cf6c93f4d7b1a067474d38a1de5641d4bd"} Nov 26 16:57:06 crc kubenswrapper[4956]: I1126 16:57:06.784162 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mvl9d" event={"ID":"ab83ee79-8dfe-4cd9-b104-49ac79683da9","Type":"ContainerStarted","Data":"9ff2b6eddb8bca418772502f989df3cca869e214836acb4f81987319928b1b8f"} Nov 26 16:57:06 crc kubenswrapper[4956]: I1126 16:57:06.800374 4956 generic.go:334] "Generic (PLEG): container finished" podID="abad77b8-1ee5-455d-9510-f76fbc9f3014" containerID="38c63208c0b4064caebb754d6514c5e54a5b68322086b869f66f4d5a3e82c725" exitCode=0 Nov 26 16:57:06 crc kubenswrapper[4956]: I1126 16:57:06.800437 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6bhv" event={"ID":"abad77b8-1ee5-455d-9510-f76fbc9f3014","Type":"ContainerDied","Data":"38c63208c0b4064caebb754d6514c5e54a5b68322086b869f66f4d5a3e82c725"} Nov 26 16:57:06 crc kubenswrapper[4956]: I1126 16:57:06.820293 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mvl9d" podStartSLOduration=2.6840675750000003 podStartE2EDuration="6.820267657s" podCreationTimestamp="2025-11-26 16:57:00 +0000 UTC" firstStartedPulling="2025-11-26 16:57:01.726885736 +0000 UTC m=+267.422846288" lastFinishedPulling="2025-11-26 16:57:05.863085808 +0000 UTC m=+271.559046370" observedRunningTime="2025-11-26 16:57:06.816175659 +0000 UTC m=+272.512136211" watchObservedRunningTime="2025-11-26 16:57:06.820267657 +0000 UTC m=+272.516228209" Nov 26 16:57:07 crc kubenswrapper[4956]: I1126 16:57:07.808487 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-c6bhv" event={"ID":"abad77b8-1ee5-455d-9510-f76fbc9f3014","Type":"ContainerStarted","Data":"7ecc978ba016829fac6030e787b495339ac1dfdcf3e91eef671fb9719d0344ce"} Nov 26 16:57:07 crc kubenswrapper[4956]: I1126 16:57:07.813409 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerStarted","Data":"1398981dbf3f4bdf02ce5fd80fb0738e1d1b5debeb1ee1683390cc84a6dcf374"} Nov 26 16:57:07 crc kubenswrapper[4956]: I1126 16:57:07.838927 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-c6bhv" podStartSLOduration=2.20802019 podStartE2EDuration="4.838900853s" podCreationTimestamp="2025-11-26 16:57:03 +0000 UTC" firstStartedPulling="2025-11-26 16:57:04.767490646 +0000 UTC m=+270.463451208" lastFinishedPulling="2025-11-26 16:57:07.398371319 +0000 UTC m=+273.094331871" observedRunningTime="2025-11-26 16:57:07.835750982 +0000 UTC m=+273.531711544" watchObservedRunningTime="2025-11-26 16:57:07.838900853 +0000 UTC m=+273.534861405" Nov 26 16:57:07 crc kubenswrapper[4956]: I1126 16:57:07.857087 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dhxg8" podStartSLOduration=2.880115585 podStartE2EDuration="6.857062485s" podCreationTimestamp="2025-11-26 16:57:01 +0000 UTC" firstStartedPulling="2025-11-26 16:57:02.73569048 +0000 UTC m=+268.431651032" lastFinishedPulling="2025-11-26 16:57:06.71263738 +0000 UTC m=+272.408597932" observedRunningTime="2025-11-26 16:57:07.854570164 +0000 UTC m=+273.550530726" watchObservedRunningTime="2025-11-26 16:57:07.857062485 +0000 UTC m=+273.553023037" Nov 26 16:57:09 crc kubenswrapper[4956]: I1126 16:57:09.206989 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:57:09 crc kubenswrapper[4956]: I1126 16:57:09.207439 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:57:09 crc kubenswrapper[4956]: I1126 16:57:09.254839 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:57:09 crc kubenswrapper[4956]: I1126 16:57:09.877300 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-kspnz" Nov 26 16:57:11 crc kubenswrapper[4956]: I1126 16:57:11.010136 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:11 crc kubenswrapper[4956]: I1126 16:57:11.010206 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:11 crc kubenswrapper[4956]: I1126 16:57:11.606495 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:11 crc kubenswrapper[4956]: I1126 16:57:11.607097 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:11 crc kubenswrapper[4956]: I1126 16:57:11.655639 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:11 crc kubenswrapper[4956]: I1126 16:57:11.889296 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 16:57:12 crc kubenswrapper[4956]: I1126 16:57:12.055629 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mvl9d" podUID="ab83ee79-8dfe-4cd9-b104-49ac79683da9" containerName="registry-server" probeResult="failure" output=< Nov 26 16:57:12 crc kubenswrapper[4956]: timeout: failed to connect service ":50051" within 1s Nov 26 16:57:12 crc kubenswrapper[4956]: > Nov 26 16:57:13 crc kubenswrapper[4956]: I1126 16:57:13.414963 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:13 crc kubenswrapper[4956]: I1126 16:57:13.415277 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:13 crc kubenswrapper[4956]: I1126 16:57:13.460469 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:13 crc kubenswrapper[4956]: I1126 16:57:13.898026 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-c6bhv" Nov 26 16:57:21 crc kubenswrapper[4956]: I1126 16:57:21.082015 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:57:21 crc kubenswrapper[4956]: I1126 16:57:21.156227 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mvl9d" Nov 26 16:58:39 crc kubenswrapper[4956]: I1126 16:58:39.552158 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 16:58:39 crc kubenswrapper[4956]: I1126 16:58:39.553279 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 16:59:09 crc kubenswrapper[4956]: I1126 16:59:09.551266 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 16:59:09 crc kubenswrapper[4956]: I1126 16:59:09.552505 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.551487 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.552420 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.552489 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.553236 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"886bfbf287a7f74a3eeafb773a2ec315c022875497b66a1ecbc8a4ff191c22b3"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.553313 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://886bfbf287a7f74a3eeafb773a2ec315c022875497b66a1ecbc8a4ff191c22b3" gracePeriod=600 Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.955551 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="886bfbf287a7f74a3eeafb773a2ec315c022875497b66a1ecbc8a4ff191c22b3" exitCode=0 Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.955710 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"886bfbf287a7f74a3eeafb773a2ec315c022875497b66a1ecbc8a4ff191c22b3"} Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.956206 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"abea58a965d35521fe5183bcbe17e6d6a5bcb02b082905c716d7ab2a2436c54e"} Nov 26 16:59:39 crc kubenswrapper[4956]: I1126 16:59:39.956254 4956 scope.go:117] "RemoveContainer" containerID="89cffde4fb042d2d8f0cb9019f281c1e1b854439e2c05daa8ce09e416a36d6cf" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.163752 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd"] Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.165610 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.168973 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.169042 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.181569 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd"] Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.274077 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nphgr\" (UniqueName: \"kubernetes.io/projected/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-kube-api-access-nphgr\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.274145 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-config-volume\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.274210 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-secret-volume\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.375609 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-secret-volume\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.375751 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nphgr\" (UniqueName: \"kubernetes.io/projected/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-kube-api-access-nphgr\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.375832 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-config-volume\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.377600 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-config-volume\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.389514 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-secret-volume\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.401399 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nphgr\" (UniqueName: \"kubernetes.io/projected/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-kube-api-access-nphgr\") pod \"collect-profiles-29402940-lg2dd\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.493121 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:00 crc kubenswrapper[4956]: I1126 17:00:00.741203 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd"] Nov 26 17:00:01 crc kubenswrapper[4956]: I1126 17:00:01.123992 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" event={"ID":"e5caa2ff-9f19-4ffb-9126-14527f32f8d3","Type":"ContainerStarted","Data":"378ab4820524c61df67b8e4d08d3c327ae982012aa609f502e9e9f5a5d9aa338"} Nov 26 17:00:02 crc kubenswrapper[4956]: I1126 17:00:02.131847 4956 generic.go:334] "Generic (PLEG): container finished" podID="e5caa2ff-9f19-4ffb-9126-14527f32f8d3" containerID="e5a2de79f2fc2b19d66bb820f842e464b7cce619acdb54421bf9adea90654e1a" exitCode=0 Nov 26 17:00:02 crc kubenswrapper[4956]: I1126 17:00:02.131955 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" event={"ID":"e5caa2ff-9f19-4ffb-9126-14527f32f8d3","Type":"ContainerDied","Data":"e5a2de79f2fc2b19d66bb820f842e464b7cce619acdb54421bf9adea90654e1a"} Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.500848 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.628743 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nphgr\" (UniqueName: \"kubernetes.io/projected/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-kube-api-access-nphgr\") pod \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.628939 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-secret-volume\") pod \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.630047 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-config-volume\") pod \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\" (UID: \"e5caa2ff-9f19-4ffb-9126-14527f32f8d3\") " Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.630663 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-config-volume" (OuterVolumeSpecName: "config-volume") pod "e5caa2ff-9f19-4ffb-9126-14527f32f8d3" (UID: "e5caa2ff-9f19-4ffb-9126-14527f32f8d3"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.636416 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "e5caa2ff-9f19-4ffb-9126-14527f32f8d3" (UID: "e5caa2ff-9f19-4ffb-9126-14527f32f8d3"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.636402 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-kube-api-access-nphgr" (OuterVolumeSpecName: "kube-api-access-nphgr") pod "e5caa2ff-9f19-4ffb-9126-14527f32f8d3" (UID: "e5caa2ff-9f19-4ffb-9126-14527f32f8d3"). InnerVolumeSpecName "kube-api-access-nphgr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.732001 4956 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.732469 4956 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 17:00:03 crc kubenswrapper[4956]: I1126 17:00:03.732493 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nphgr\" (UniqueName: \"kubernetes.io/projected/e5caa2ff-9f19-4ffb-9126-14527f32f8d3-kube-api-access-nphgr\") on node \"crc\" DevicePath \"\"" Nov 26 17:00:04 crc kubenswrapper[4956]: I1126 17:00:04.154828 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" event={"ID":"e5caa2ff-9f19-4ffb-9126-14527f32f8d3","Type":"ContainerDied","Data":"378ab4820524c61df67b8e4d08d3c327ae982012aa609f502e9e9f5a5d9aa338"} Nov 26 17:00:04 crc kubenswrapper[4956]: I1126 17:00:04.154931 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="378ab4820524c61df67b8e4d08d3c327ae982012aa609f502e9e9f5a5d9aa338" Nov 26 17:00:04 crc kubenswrapper[4956]: I1126 17:00:04.154998 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402940-lg2dd" Nov 26 17:01:10 crc kubenswrapper[4956]: I1126 17:01:10.938974 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9sbpj"] Nov 26 17:01:10 crc kubenswrapper[4956]: E1126 17:01:10.940245 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5caa2ff-9f19-4ffb-9126-14527f32f8d3" containerName="collect-profiles" Nov 26 17:01:10 crc kubenswrapper[4956]: I1126 17:01:10.940266 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5caa2ff-9f19-4ffb-9126-14527f32f8d3" containerName="collect-profiles" Nov 26 17:01:10 crc kubenswrapper[4956]: I1126 17:01:10.940398 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5caa2ff-9f19-4ffb-9126-14527f32f8d3" containerName="collect-profiles" Nov 26 17:01:10 crc kubenswrapper[4956]: I1126 17:01:10.941050 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:10 crc kubenswrapper[4956]: I1126 17:01:10.958331 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9sbpj"] Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102026 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28fcd3d2-c616-4867-84a0-29f04287931e-registry-certificates\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102208 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28fcd3d2-c616-4867-84a0-29f04287931e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102423 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-registry-tls\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102580 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28fcd3d2-c616-4867-84a0-29f04287931e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102625 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28fcd3d2-c616-4867-84a0-29f04287931e-trusted-ca\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102660 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rdv9j\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-kube-api-access-rdv9j\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102691 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-bound-sa-token\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.102754 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.134196 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204494 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28fcd3d2-c616-4867-84a0-29f04287931e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204548 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28fcd3d2-c616-4867-84a0-29f04287931e-trusted-ca\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204566 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdv9j\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-kube-api-access-rdv9j\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204586 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-bound-sa-token\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204629 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28fcd3d2-c616-4867-84a0-29f04287931e-registry-certificates\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204658 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28fcd3d2-c616-4867-84a0-29f04287931e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.204688 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-registry-tls\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.205784 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/28fcd3d2-c616-4867-84a0-29f04287931e-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.207558 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/28fcd3d2-c616-4867-84a0-29f04287931e-registry-certificates\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.208266 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/28fcd3d2-c616-4867-84a0-29f04287931e-trusted-ca\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.212245 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/28fcd3d2-c616-4867-84a0-29f04287931e-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.213910 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-registry-tls\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.223334 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdv9j\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-kube-api-access-rdv9j\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.235719 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/28fcd3d2-c616-4867-84a0-29f04287931e-bound-sa-token\") pod \"image-registry-66df7c8f76-9sbpj\" (UID: \"28fcd3d2-c616-4867-84a0-29f04287931e\") " pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.266333 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:11 crc kubenswrapper[4956]: I1126 17:01:11.552305 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9sbpj"] Nov 26 17:01:12 crc kubenswrapper[4956]: I1126 17:01:12.041534 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" event={"ID":"28fcd3d2-c616-4867-84a0-29f04287931e","Type":"ContainerStarted","Data":"3144e3535dcce57a853622aeaa32032563e5da8b5feac96b544aaac3ad17f3a9"} Nov 26 17:01:12 crc kubenswrapper[4956]: I1126 17:01:12.041603 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" event={"ID":"28fcd3d2-c616-4867-84a0-29f04287931e","Type":"ContainerStarted","Data":"a226723d98dcb33a51b1a9ff093775e47239313188760cfe0bac7b3bf31da01e"} Nov 26 17:01:12 crc kubenswrapper[4956]: I1126 17:01:12.041769 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:12 crc kubenswrapper[4956]: I1126 17:01:12.074809 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" podStartSLOduration=2.074780708 podStartE2EDuration="2.074780708s" podCreationTimestamp="2025-11-26 17:01:10 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:01:12.068834049 +0000 UTC m=+517.764794681" watchObservedRunningTime="2025-11-26 17:01:12.074780708 +0000 UTC m=+517.770741270" Nov 26 17:01:31 crc kubenswrapper[4956]: I1126 17:01:31.271372 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9sbpj" Nov 26 17:01:31 crc kubenswrapper[4956]: I1126 17:01:31.329279 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xjxmw"] Nov 26 17:01:39 crc kubenswrapper[4956]: I1126 17:01:39.551734 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:01:39 crc kubenswrapper[4956]: I1126 17:01:39.552530 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:01:56 crc kubenswrapper[4956]: I1126 17:01:56.375590 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" podUID="8394076a-e17a-4cea-a904-eb269c237f00" containerName="registry" containerID="cri-o://cd18ff7527c1e635622e62921b7abd26b075602ccdb89553a273b90078208974" gracePeriod=30 Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.382530 4956 generic.go:334] "Generic (PLEG): container finished" podID="8394076a-e17a-4cea-a904-eb269c237f00" containerID="cd18ff7527c1e635622e62921b7abd26b075602ccdb89553a273b90078208974" exitCode=0 Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.383229 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" event={"ID":"8394076a-e17a-4cea-a904-eb269c237f00","Type":"ContainerDied","Data":"cd18ff7527c1e635622e62921b7abd26b075602ccdb89553a273b90078208974"} Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.383345 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" event={"ID":"8394076a-e17a-4cea-a904-eb269c237f00","Type":"ContainerDied","Data":"d328275e411872b8ef64c2a4e33fc2eec646ca0c742824ea6bd3b974b4338bd2"} Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.383376 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d328275e411872b8ef64c2a4e33fc2eec646ca0c742824ea6bd3b974b4338bd2" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.396373 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479533 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-registry-certificates\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479666 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8394076a-e17a-4cea-a904-eb269c237f00-ca-trust-extracted\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479701 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-trusted-ca\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479737 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-55rk5\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-kube-api-access-55rk5\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479762 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-registry-tls\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479798 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8394076a-e17a-4cea-a904-eb269c237f00-installation-pull-secrets\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.479844 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-bound-sa-token\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.480148 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8394076a-e17a-4cea-a904-eb269c237f00\" (UID: \"8394076a-e17a-4cea-a904-eb269c237f00\") " Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.481240 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.481338 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.488832 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.489267 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8394076a-e17a-4cea-a904-eb269c237f00-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.489538 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.493301 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-kube-api-access-55rk5" (OuterVolumeSpecName: "kube-api-access-55rk5") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "kube-api-access-55rk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.505408 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8394076a-e17a-4cea-a904-eb269c237f00-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.527531 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "8394076a-e17a-4cea-a904-eb269c237f00" (UID: "8394076a-e17a-4cea-a904-eb269c237f00"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582698 4956 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-registry-certificates\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582761 4956 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8394076a-e17a-4cea-a904-eb269c237f00-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582783 4956 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8394076a-e17a-4cea-a904-eb269c237f00-trusted-ca\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582802 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-55rk5\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-kube-api-access-55rk5\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582821 4956 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-registry-tls\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582843 4956 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8394076a-e17a-4cea-a904-eb269c237f00-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:57 crc kubenswrapper[4956]: I1126 17:01:57.582860 4956 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8394076a-e17a-4cea-a904-eb269c237f00-bound-sa-token\") on node \"crc\" DevicePath \"\"" Nov 26 17:01:58 crc kubenswrapper[4956]: I1126 17:01:58.389396 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-xjxmw" Nov 26 17:01:58 crc kubenswrapper[4956]: I1126 17:01:58.437875 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xjxmw"] Nov 26 17:01:58 crc kubenswrapper[4956]: I1126 17:01:58.440633 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-xjxmw"] Nov 26 17:01:59 crc kubenswrapper[4956]: I1126 17:01:59.005390 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8394076a-e17a-4cea-a904-eb269c237f00" path="/var/lib/kubelet/pods/8394076a-e17a-4cea-a904-eb269c237f00/volumes" Nov 26 17:02:09 crc kubenswrapper[4956]: I1126 17:02:09.551912 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:02:09 crc kubenswrapper[4956]: I1126 17:02:09.552999 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:02:35 crc kubenswrapper[4956]: I1126 17:02:35.292712 4956 scope.go:117] "RemoveContainer" containerID="cd18ff7527c1e635622e62921b7abd26b075602ccdb89553a273b90078208974" Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.551999 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.552555 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.552639 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.553770 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"abea58a965d35521fe5183bcbe17e6d6a5bcb02b082905c716d7ab2a2436c54e"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.553912 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://abea58a965d35521fe5183bcbe17e6d6a5bcb02b082905c716d7ab2a2436c54e" gracePeriod=600 Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.732785 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="abea58a965d35521fe5183bcbe17e6d6a5bcb02b082905c716d7ab2a2436c54e" exitCode=0 Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.732901 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"abea58a965d35521fe5183bcbe17e6d6a5bcb02b082905c716d7ab2a2436c54e"} Nov 26 17:02:39 crc kubenswrapper[4956]: I1126 17:02:39.733243 4956 scope.go:117] "RemoveContainer" containerID="886bfbf287a7f74a3eeafb773a2ec315c022875497b66a1ecbc8a4ff191c22b3" Nov 26 17:02:40 crc kubenswrapper[4956]: I1126 17:02:40.743536 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"60fb36ae8e9dbdf5a9ddd39194e72f5c72fcc2005c642dd846c7d001b368d385"} Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.894662 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f2thc"] Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.895718 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-controller" containerID="cri-o://5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.895768 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="nbdb" containerID="cri-o://2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.895889 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="northd" containerID="cri-o://b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.895942 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.895971 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-node" containerID="cri-o://c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.896003 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-acl-logging" containerID="cri-o://9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.896180 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="sbdb" containerID="cri-o://c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" gracePeriod=30 Nov 26 17:02:56 crc kubenswrapper[4956]: I1126 17:02:56.936412 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" containerID="cri-o://324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" gracePeriod=30 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.297877 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/3.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.301039 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovn-acl-logging/0.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.301506 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovn-controller/0.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.301900 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358500 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-w2dsx"] Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358831 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-acl-logging" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358849 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-acl-logging" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358863 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-node" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358888 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-node" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358899 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8394076a-e17a-4cea-a904-eb269c237f00" containerName="registry" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358906 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8394076a-e17a-4cea-a904-eb269c237f00" containerName="registry" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358917 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358927 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358941 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358949 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358961 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="nbdb" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358969 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="nbdb" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358981 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kubecfg-setup" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.358988 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kubecfg-setup" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.358998 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359011 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.359020 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359026 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.359036 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359043 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.359053 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="sbdb" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359060 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="sbdb" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.359071 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="northd" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359079 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="northd" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.359092 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359099 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359220 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-node" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359235 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="kube-rbac-proxy-ovn-metrics" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359244 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="northd" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359256 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359264 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="nbdb" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359277 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359288 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-acl-logging" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359299 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovn-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359309 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359319 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="sbdb" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359326 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8394076a-e17a-4cea-a904-eb269c237f00" containerName="registry" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359337 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.359452 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359461 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.359572 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" containerName="ovnkube-controller" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.361683 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.478924 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-bin\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479092 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-netns\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479127 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-systemd-units\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479152 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-slash\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479176 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-ovn-kubernetes\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479208 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-var-lib-cni-networks-ovn-kubernetes\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479259 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-script-lib\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479307 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-ovn\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479330 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-log-socket\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479351 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-kubelet\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479380 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-config\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479409 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-env-overrides\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479435 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-node-log\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479458 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zzdlc\" (UniqueName: \"kubernetes.io/projected/46bd4702-b565-4452-be2c-36a3f1f48621-kube-api-access-zzdlc\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479483 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-netd\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479512 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-systemd\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479532 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46bd4702-b565-4452-be2c-36a3f1f48621-ovn-node-metrics-cert\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479559 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-etc-openvswitch\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479582 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-openvswitch\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479598 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-var-lib-openvswitch\") pod \"46bd4702-b565-4452-be2c-36a3f1f48621\" (UID: \"46bd4702-b565-4452-be2c-36a3f1f48621\") " Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479800 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-slash\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479829 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-run-netns\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479849 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479901 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gbn2b\" (UniqueName: \"kubernetes.io/projected/2130a868-de00-4ede-9423-941a66da84e0-kube-api-access-gbn2b\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479928 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-cni-netd\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479951 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-log-socket\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479996 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-etc-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480028 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-ovnkube-config\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480058 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-node-log\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480087 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-cni-bin\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480107 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-kubelet\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480134 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-var-lib-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480152 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-systemd\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480179 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480208 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-run-ovn-kubernetes\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480231 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-env-overrides\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480252 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2130a868-de00-4ede-9423-941a66da84e0-ovn-node-metrics-cert\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480270 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-ovnkube-script-lib\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480294 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-systemd-units\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480312 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-ovn\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.479041 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480395 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480415 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480433 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-slash" (OuterVolumeSpecName: "host-slash") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480450 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480470 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480819 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480857 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-log-socket" (OuterVolumeSpecName: "log-socket") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.480915 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.481230 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-node-log" (OuterVolumeSpecName: "node-log") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.481264 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.481302 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.481323 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.481358 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.482071 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.482130 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.482290 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.488253 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/46bd4702-b565-4452-be2c-36a3f1f48621-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.489937 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46bd4702-b565-4452-be2c-36a3f1f48621-kube-api-access-zzdlc" (OuterVolumeSpecName: "kube-api-access-zzdlc") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "kube-api-access-zzdlc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.497954 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "46bd4702-b565-4452-be2c-36a3f1f48621" (UID: "46bd4702-b565-4452-be2c-36a3f1f48621"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.581996 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-var-lib-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582062 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-systemd\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582092 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582116 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-run-ovn-kubernetes\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582134 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-env-overrides\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582150 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2130a868-de00-4ede-9423-941a66da84e0-ovn-node-metrics-cert\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582164 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-ovnkube-script-lib\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582179 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-systemd-units\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582196 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-ovn\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582213 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-slash\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582229 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-run-netns\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582242 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582265 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gbn2b\" (UniqueName: \"kubernetes.io/projected/2130a868-de00-4ede-9423-941a66da84e0-kube-api-access-gbn2b\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582285 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-cni-netd\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582308 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-log-socket\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582341 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-etc-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582362 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-ovnkube-config\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582381 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-node-log\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582516 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-ovn\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582632 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-run-ovn-kubernetes\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582685 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-slash\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582648 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-run-netns\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582769 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-cni-netd\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582832 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-var-lib-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582858 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-etc-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582898 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-openvswitch\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582927 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-systemd-units\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582951 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582988 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-run-systemd\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.583020 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-node-log\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.583579 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-ovnkube-config\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.583667 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-ovnkube-script-lib\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.582635 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-log-socket\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584123 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/2130a868-de00-4ede-9423-941a66da84e0-env-overrides\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584203 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-cni-bin\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584225 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-kubelet\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584280 4956 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-systemd\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584294 4956 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/46bd4702-b565-4452-be2c-36a3f1f48621-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584307 4956 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584319 4956 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584334 4956 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584347 4956 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-bin\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584357 4956 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-netns\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584369 4956 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-systemd-units\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584378 4956 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-slash\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584387 4956 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584398 4956 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584409 4956 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584419 4956 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-run-ovn\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584428 4956 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-log-socket\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584439 4956 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-kubelet\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584449 4956 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-ovnkube-config\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584459 4956 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/46bd4702-b565-4452-be2c-36a3f1f48621-env-overrides\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584470 4956 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-node-log\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584482 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zzdlc\" (UniqueName: \"kubernetes.io/projected/46bd4702-b565-4452-be2c-36a3f1f48621-kube-api-access-zzdlc\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584494 4956 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/46bd4702-b565-4452-be2c-36a3f1f48621-host-cni-netd\") on node \"crc\" DevicePath \"\"" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584522 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-kubelet\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.584552 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/2130a868-de00-4ede-9423-941a66da84e0-host-cni-bin\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.587345 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/2130a868-de00-4ede-9423-941a66da84e0-ovn-node-metrics-cert\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.602120 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gbn2b\" (UniqueName: \"kubernetes.io/projected/2130a868-de00-4ede-9423-941a66da84e0-kube-api-access-gbn2b\") pod \"ovnkube-node-w2dsx\" (UID: \"2130a868-de00-4ede-9423-941a66da84e0\") " pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.683368 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:02:57 crc kubenswrapper[4956]: W1126 17:02:57.712913 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2130a868_de00_4ede_9423_941a66da84e0.slice/crio-183b1fcc98673a4c1c47e5921ca2289870182120bdbb539cdc4d41fe6b0f8d0b WatchSource:0}: Error finding container 183b1fcc98673a4c1c47e5921ca2289870182120bdbb539cdc4d41fe6b0f8d0b: Status 404 returned error can't find the container with id 183b1fcc98673a4c1c47e5921ca2289870182120bdbb539cdc4d41fe6b0f8d0b Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.876941 4956 generic.go:334] "Generic (PLEG): container finished" podID="2130a868-de00-4ede-9423-941a66da84e0" containerID="ae2751ba8a4b448c8396c781cae5ceea4511f03ae17a33a06a02fc48e7c2891c" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.877461 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerDied","Data":"ae2751ba8a4b448c8396c781cae5ceea4511f03ae17a33a06a02fc48e7c2891c"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.877513 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"183b1fcc98673a4c1c47e5921ca2289870182120bdbb539cdc4d41fe6b0f8d0b"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.885425 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovnkube-controller/3.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.889121 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovn-acl-logging/0.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.889597 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f2thc_46bd4702-b565-4452-be2c-36a3f1f48621/ovn-controller/0.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890428 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890470 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890483 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890495 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890504 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890515 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" exitCode=0 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890524 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" exitCode=143 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890534 4956 generic.go:334] "Generic (PLEG): container finished" podID="46bd4702-b565-4452-be2c-36a3f1f48621" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" exitCode=143 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890517 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890641 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890662 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890682 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890696 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890709 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890721 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890738 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890754 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890762 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890771 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890779 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890787 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890794 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890802 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890809 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890819 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890744 4956 scope.go:117] "RemoveContainer" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.890830 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891100 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891122 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891150 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891159 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891170 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891179 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891189 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891200 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891211 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891240 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891268 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891279 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891288 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891298 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891307 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891316 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891325 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891335 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891344 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891352 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891366 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f2thc" event={"ID":"46bd4702-b565-4452-be2c-36a3f1f48621","Type":"ContainerDied","Data":"7ffa3211be69b02f996dc04517ad911040022ba015618c443c16d8e2b7a4e871"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891381 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891396 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891406 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891415 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891425 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891434 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891443 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891452 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891463 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.891472 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.892977 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/2.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.893499 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/1.log" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.893550 4956 generic.go:334] "Generic (PLEG): container finished" podID="80ac0050-7f55-4845-ace7-a89f2442f884" containerID="43e4c0b34591d5c93dae72d209597bd85a17d5098e385aafdcdf66e667e47b4a" exitCode=2 Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.893611 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerDied","Data":"43e4c0b34591d5c93dae72d209597bd85a17d5098e385aafdcdf66e667e47b4a"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.893657 4956 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb"} Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.894305 4956 scope.go:117] "RemoveContainer" containerID="43e4c0b34591d5c93dae72d209597bd85a17d5098e385aafdcdf66e667e47b4a" Nov 26 17:02:57 crc kubenswrapper[4956]: E1126 17:02:57.894531 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m4dm4_openshift-multus(80ac0050-7f55-4845-ace7-a89f2442f884)\"" pod="openshift-multus/multus-m4dm4" podUID="80ac0050-7f55-4845-ace7-a89f2442f884" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.933451 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.959271 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f2thc"] Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.961915 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f2thc"] Nov 26 17:02:57 crc kubenswrapper[4956]: I1126 17:02:57.967005 4956 scope.go:117] "RemoveContainer" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.001157 4956 scope.go:117] "RemoveContainer" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.021578 4956 scope.go:117] "RemoveContainer" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.059598 4956 scope.go:117] "RemoveContainer" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.081283 4956 scope.go:117] "RemoveContainer" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.107658 4956 scope.go:117] "RemoveContainer" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.134313 4956 scope.go:117] "RemoveContainer" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.152986 4956 scope.go:117] "RemoveContainer" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.175056 4956 scope.go:117] "RemoveContainer" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.178016 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": container with ID starting with 324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5 not found: ID does not exist" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.178073 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} err="failed to get container status \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": rpc error: code = NotFound desc = could not find container \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": container with ID starting with 324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.178107 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.179765 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": container with ID starting with 0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae not found: ID does not exist" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.179823 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} err="failed to get container status \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": rpc error: code = NotFound desc = could not find container \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": container with ID starting with 0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.179878 4956 scope.go:117] "RemoveContainer" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.181876 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": container with ID starting with c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705 not found: ID does not exist" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.181941 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} err="failed to get container status \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": rpc error: code = NotFound desc = could not find container \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": container with ID starting with c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.181980 4956 scope.go:117] "RemoveContainer" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.183124 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": container with ID starting with 2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756 not found: ID does not exist" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.183161 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} err="failed to get container status \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": rpc error: code = NotFound desc = could not find container \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": container with ID starting with 2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.183187 4956 scope.go:117] "RemoveContainer" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.193308 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": container with ID starting with b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166 not found: ID does not exist" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.193359 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} err="failed to get container status \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": rpc error: code = NotFound desc = could not find container \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": container with ID starting with b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.193392 4956 scope.go:117] "RemoveContainer" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.194387 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": container with ID starting with 8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd not found: ID does not exist" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.194413 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} err="failed to get container status \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": rpc error: code = NotFound desc = could not find container \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": container with ID starting with 8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.194430 4956 scope.go:117] "RemoveContainer" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.194708 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": container with ID starting with c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8 not found: ID does not exist" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.194729 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} err="failed to get container status \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": rpc error: code = NotFound desc = could not find container \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": container with ID starting with c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.194742 4956 scope.go:117] "RemoveContainer" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.195580 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": container with ID starting with 9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517 not found: ID does not exist" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.195608 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} err="failed to get container status \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": rpc error: code = NotFound desc = could not find container \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": container with ID starting with 9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.195627 4956 scope.go:117] "RemoveContainer" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.196075 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": container with ID starting with 5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c not found: ID does not exist" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.196097 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} err="failed to get container status \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": rpc error: code = NotFound desc = could not find container \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": container with ID starting with 5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.196116 4956 scope.go:117] "RemoveContainer" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" Nov 26 17:02:58 crc kubenswrapper[4956]: E1126 17:02:58.196467 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": container with ID starting with 00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221 not found: ID does not exist" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.196523 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} err="failed to get container status \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": rpc error: code = NotFound desc = could not find container \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": container with ID starting with 00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.196561 4956 scope.go:117] "RemoveContainer" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.196859 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} err="failed to get container status \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": rpc error: code = NotFound desc = could not find container \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": container with ID starting with 324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.196897 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.197270 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} err="failed to get container status \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": rpc error: code = NotFound desc = could not find container \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": container with ID starting with 0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.197326 4956 scope.go:117] "RemoveContainer" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.197676 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} err="failed to get container status \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": rpc error: code = NotFound desc = could not find container \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": container with ID starting with c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.197702 4956 scope.go:117] "RemoveContainer" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.197959 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} err="failed to get container status \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": rpc error: code = NotFound desc = could not find container \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": container with ID starting with 2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.197982 4956 scope.go:117] "RemoveContainer" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.198228 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} err="failed to get container status \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": rpc error: code = NotFound desc = could not find container \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": container with ID starting with b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.198259 4956 scope.go:117] "RemoveContainer" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.198498 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} err="failed to get container status \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": rpc error: code = NotFound desc = could not find container \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": container with ID starting with 8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.198523 4956 scope.go:117] "RemoveContainer" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.198749 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} err="failed to get container status \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": rpc error: code = NotFound desc = could not find container \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": container with ID starting with c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.198770 4956 scope.go:117] "RemoveContainer" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.199025 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} err="failed to get container status \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": rpc error: code = NotFound desc = could not find container \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": container with ID starting with 9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.199049 4956 scope.go:117] "RemoveContainer" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.199715 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} err="failed to get container status \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": rpc error: code = NotFound desc = could not find container \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": container with ID starting with 5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.199742 4956 scope.go:117] "RemoveContainer" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.200109 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} err="failed to get container status \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": rpc error: code = NotFound desc = could not find container \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": container with ID starting with 00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.200134 4956 scope.go:117] "RemoveContainer" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.200423 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} err="failed to get container status \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": rpc error: code = NotFound desc = could not find container \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": container with ID starting with 324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.200462 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.200787 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} err="failed to get container status \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": rpc error: code = NotFound desc = could not find container \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": container with ID starting with 0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.200811 4956 scope.go:117] "RemoveContainer" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.201108 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} err="failed to get container status \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": rpc error: code = NotFound desc = could not find container \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": container with ID starting with c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.201133 4956 scope.go:117] "RemoveContainer" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.201422 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} err="failed to get container status \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": rpc error: code = NotFound desc = could not find container \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": container with ID starting with 2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.201455 4956 scope.go:117] "RemoveContainer" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.201721 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} err="failed to get container status \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": rpc error: code = NotFound desc = could not find container \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": container with ID starting with b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.201748 4956 scope.go:117] "RemoveContainer" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.202051 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} err="failed to get container status \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": rpc error: code = NotFound desc = could not find container \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": container with ID starting with 8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.202074 4956 scope.go:117] "RemoveContainer" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.202337 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} err="failed to get container status \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": rpc error: code = NotFound desc = could not find container \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": container with ID starting with c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.202356 4956 scope.go:117] "RemoveContainer" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.202654 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} err="failed to get container status \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": rpc error: code = NotFound desc = could not find container \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": container with ID starting with 9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.202689 4956 scope.go:117] "RemoveContainer" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.203258 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} err="failed to get container status \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": rpc error: code = NotFound desc = could not find container \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": container with ID starting with 5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.203286 4956 scope.go:117] "RemoveContainer" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.203553 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} err="failed to get container status \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": rpc error: code = NotFound desc = could not find container \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": container with ID starting with 00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.203586 4956 scope.go:117] "RemoveContainer" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.203887 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} err="failed to get container status \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": rpc error: code = NotFound desc = could not find container \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": container with ID starting with 324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.203923 4956 scope.go:117] "RemoveContainer" containerID="0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.204265 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae"} err="failed to get container status \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": rpc error: code = NotFound desc = could not find container \"0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae\": container with ID starting with 0298756e8567a366c796b034735f6b1eb73247b5e4b3240a607977f2242a3fae not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.204324 4956 scope.go:117] "RemoveContainer" containerID="c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.204626 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705"} err="failed to get container status \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": rpc error: code = NotFound desc = could not find container \"c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705\": container with ID starting with c87a3c6e282e87fe62d3f711cc59dfbff548993ad63971cc37ba34079af48705 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.204660 4956 scope.go:117] "RemoveContainer" containerID="2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.205084 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756"} err="failed to get container status \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": rpc error: code = NotFound desc = could not find container \"2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756\": container with ID starting with 2f0e79feee76eca90c16c0105105d1800ffe6c67ccec46732f70db3d67e16756 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.205110 4956 scope.go:117] "RemoveContainer" containerID="b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.205495 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166"} err="failed to get container status \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": rpc error: code = NotFound desc = could not find container \"b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166\": container with ID starting with b80a4f05b9c7a0b4e2ebafa266235f9563b4bd7cde3e8de499414eb8635d8166 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.205523 4956 scope.go:117] "RemoveContainer" containerID="8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.205821 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd"} err="failed to get container status \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": rpc error: code = NotFound desc = could not find container \"8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd\": container with ID starting with 8b8a7ab43e7f14a7fa71706007b8746b144390dc0116f7690703d38cae0159dd not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.205841 4956 scope.go:117] "RemoveContainer" containerID="c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.206214 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8"} err="failed to get container status \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": rpc error: code = NotFound desc = could not find container \"c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8\": container with ID starting with c7ea0d1a12beba94109849e559ba26ae155e79a21efbc4b92bf8369b4165fdd8 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.206241 4956 scope.go:117] "RemoveContainer" containerID="9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.206535 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517"} err="failed to get container status \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": rpc error: code = NotFound desc = could not find container \"9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517\": container with ID starting with 9d68b0914f83ba8d5128551ff8c8cebdc791e8349a07696844c57b7e90f0d517 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.206580 4956 scope.go:117] "RemoveContainer" containerID="5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.206884 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c"} err="failed to get container status \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": rpc error: code = NotFound desc = could not find container \"5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c\": container with ID starting with 5169f9ac03cc3b3e3f25a2d1df58bc3a1164a954091c62d5e62fd7a7002d391c not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.206906 4956 scope.go:117] "RemoveContainer" containerID="00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.207105 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221"} err="failed to get container status \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": rpc error: code = NotFound desc = could not find container \"00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221\": container with ID starting with 00409c5342cc1f95df09f8bbdba5765ab81c979dbce0e2c64f184e4d4a1ab221 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.207123 4956 scope.go:117] "RemoveContainer" containerID="324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.207488 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5"} err="failed to get container status \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": rpc error: code = NotFound desc = could not find container \"324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5\": container with ID starting with 324a9745f378967072be7e2e38a4a78f906ffb82428c716fdb64bbc9ea7172e5 not found: ID does not exist" Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.903107 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"15341333a1ddcce9ca8dea4024d3b08fb6695db02a5525d542bf4791bb84ee11"} Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.903626 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"52d82e28a8b1b85eccca7490311e0c8f0768b2d7620a50a3fce45e2a21c241ef"} Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.903640 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"d720fa524f0700c6ea44abe27745591a42a759611f64cfb8211b8dbb94ee0f7b"} Nov 26 17:02:58 crc kubenswrapper[4956]: I1126 17:02:58.903652 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"51770602320fd5e20545a084fbe51b94f9f066b6038b216d7f23e0797ff55086"} Nov 26 17:02:59 crc kubenswrapper[4956]: I1126 17:02:59.003026 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="46bd4702-b565-4452-be2c-36a3f1f48621" path="/var/lib/kubelet/pods/46bd4702-b565-4452-be2c-36a3f1f48621/volumes" Nov 26 17:02:59 crc kubenswrapper[4956]: I1126 17:02:59.916673 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"3cea376b6655029e053b4a5e533c58708bf74ab0598488bf929d655063f05337"} Nov 26 17:02:59 crc kubenswrapper[4956]: I1126 17:02:59.917124 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"4f3bc610a20ab78cca93fd81fefc0d7534df91c506c380ad879d3cd7b0dee723"} Nov 26 17:03:01 crc kubenswrapper[4956]: I1126 17:03:01.936491 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"7ae12ca820a42b9a104ee766cc93e37172843070a4bf171140f385d37834bac3"} Nov 26 17:03:03 crc kubenswrapper[4956]: I1126 17:03:03.956552 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" event={"ID":"2130a868-de00-4ede-9423-941a66da84e0","Type":"ContainerStarted","Data":"38f1a026248e3adde0610f7da792e74bcff630ae7dc221fe1e26dd8a3151e9e4"} Nov 26 17:03:03 crc kubenswrapper[4956]: I1126 17:03:03.957984 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:03:03 crc kubenswrapper[4956]: I1126 17:03:03.958009 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:03:03 crc kubenswrapper[4956]: I1126 17:03:03.986964 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:03:03 crc kubenswrapper[4956]: I1126 17:03:03.991106 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" podStartSLOduration=6.99108786 podStartE2EDuration="6.99108786s" podCreationTimestamp="2025-11-26 17:02:57 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:03:03.987324673 +0000 UTC m=+629.683285225" watchObservedRunningTime="2025-11-26 17:03:03.99108786 +0000 UTC m=+629.687048402" Nov 26 17:03:04 crc kubenswrapper[4956]: I1126 17:03:04.967551 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:03:05 crc kubenswrapper[4956]: I1126 17:03:05.002759 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:03:10 crc kubenswrapper[4956]: I1126 17:03:10.995463 4956 scope.go:117] "RemoveContainer" containerID="43e4c0b34591d5c93dae72d209597bd85a17d5098e385aafdcdf66e667e47b4a" Nov 26 17:03:10 crc kubenswrapper[4956]: E1126 17:03:10.996207 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-m4dm4_openshift-multus(80ac0050-7f55-4845-ace7-a89f2442f884)\"" pod="openshift-multus/multus-m4dm4" podUID="80ac0050-7f55-4845-ace7-a89f2442f884" Nov 26 17:03:24 crc kubenswrapper[4956]: I1126 17:03:24.998111 4956 scope.go:117] "RemoveContainer" containerID="43e4c0b34591d5c93dae72d209597bd85a17d5098e385aafdcdf66e667e47b4a" Nov 26 17:03:26 crc kubenswrapper[4956]: I1126 17:03:26.124183 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/2.log" Nov 26 17:03:26 crc kubenswrapper[4956]: I1126 17:03:26.125016 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/1.log" Nov 26 17:03:26 crc kubenswrapper[4956]: I1126 17:03:26.125067 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-m4dm4" event={"ID":"80ac0050-7f55-4845-ace7-a89f2442f884","Type":"ContainerStarted","Data":"fe1add44bbe72cb15bae307c6bdc5667355e2b79e66232ef32adbb5e42ea5808"} Nov 26 17:03:27 crc kubenswrapper[4956]: I1126 17:03:27.713106 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-w2dsx" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.138097 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv"] Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.141940 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.152554 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.166952 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv"] Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.288572 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nt4wv\" (UniqueName: \"kubernetes.io/projected/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-kube-api-access-nt4wv\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.288752 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.288812 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.390622 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nt4wv\" (UniqueName: \"kubernetes.io/projected/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-kube-api-access-nt4wv\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.390799 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.390842 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.391532 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-util\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.391706 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-bundle\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.418307 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nt4wv\" (UniqueName: \"kubernetes.io/projected/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-kube-api-access-nt4wv\") pod \"e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.505399 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:30 crc kubenswrapper[4956]: I1126 17:03:30.952775 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv"] Nov 26 17:03:31 crc kubenswrapper[4956]: I1126 17:03:31.166414 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" event={"ID":"5da5ddac-6e96-4800-b178-a03b3cdc1dfa","Type":"ContainerStarted","Data":"1c024cb7b74bbdc7fb72b41f93b8859e6a2fd50ce4cfbf65741fe8b2855d626b"} Nov 26 17:03:32 crc kubenswrapper[4956]: I1126 17:03:32.173474 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" event={"ID":"5da5ddac-6e96-4800-b178-a03b3cdc1dfa","Type":"ContainerStarted","Data":"cdadee1ca89ae59dbef58462d6ffd01a66a7c39ba2c61af78e3e7381f9d71dbb"} Nov 26 17:03:33 crc kubenswrapper[4956]: I1126 17:03:33.179927 4956 generic.go:334] "Generic (PLEG): container finished" podID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerID="cdadee1ca89ae59dbef58462d6ffd01a66a7c39ba2c61af78e3e7381f9d71dbb" exitCode=0 Nov 26 17:03:33 crc kubenswrapper[4956]: I1126 17:03:33.179985 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" event={"ID":"5da5ddac-6e96-4800-b178-a03b3cdc1dfa","Type":"ContainerDied","Data":"cdadee1ca89ae59dbef58462d6ffd01a66a7c39ba2c61af78e3e7381f9d71dbb"} Nov 26 17:03:33 crc kubenswrapper[4956]: I1126 17:03:33.183839 4956 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 17:03:35 crc kubenswrapper[4956]: I1126 17:03:35.346042 4956 scope.go:117] "RemoveContainer" containerID="bbca0c199ddbc0a280796308fe838b51f10046711217ee408b656b4b1f0fd1fb" Nov 26 17:03:36 crc kubenswrapper[4956]: I1126 17:03:36.220641 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-m4dm4_80ac0050-7f55-4845-ace7-a89f2442f884/kube-multus/2.log" Nov 26 17:03:38 crc kubenswrapper[4956]: I1126 17:03:38.242028 4956 generic.go:334] "Generic (PLEG): container finished" podID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerID="608d67fc5bcbeed7157e76d04c89027d3f646061f43f1a3fab3a906bc965182e" exitCode=0 Nov 26 17:03:38 crc kubenswrapper[4956]: I1126 17:03:38.242248 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" event={"ID":"5da5ddac-6e96-4800-b178-a03b3cdc1dfa","Type":"ContainerDied","Data":"608d67fc5bcbeed7157e76d04c89027d3f646061f43f1a3fab3a906bc965182e"} Nov 26 17:03:39 crc kubenswrapper[4956]: I1126 17:03:39.254821 4956 generic.go:334] "Generic (PLEG): container finished" podID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerID="01da73d3bc3a93a4e087ff719ef54bd67d93548ad49120f99a41f8b7fb6e7f4b" exitCode=0 Nov 26 17:03:39 crc kubenswrapper[4956]: I1126 17:03:39.255002 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" event={"ID":"5da5ddac-6e96-4800-b178-a03b3cdc1dfa","Type":"ContainerDied","Data":"01da73d3bc3a93a4e087ff719ef54bd67d93548ad49120f99a41f8b7fb6e7f4b"} Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.563064 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.655471 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nt4wv\" (UniqueName: \"kubernetes.io/projected/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-kube-api-access-nt4wv\") pod \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.655553 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-util\") pod \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.655647 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-bundle\") pod \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\" (UID: \"5da5ddac-6e96-4800-b178-a03b3cdc1dfa\") " Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.656676 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-bundle" (OuterVolumeSpecName: "bundle") pod "5da5ddac-6e96-4800-b178-a03b3cdc1dfa" (UID: "5da5ddac-6e96-4800-b178-a03b3cdc1dfa"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.663956 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-kube-api-access-nt4wv" (OuterVolumeSpecName: "kube-api-access-nt4wv") pod "5da5ddac-6e96-4800-b178-a03b3cdc1dfa" (UID: "5da5ddac-6e96-4800-b178-a03b3cdc1dfa"). InnerVolumeSpecName "kube-api-access-nt4wv". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.668977 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-util" (OuterVolumeSpecName: "util") pod "5da5ddac-6e96-4800-b178-a03b3cdc1dfa" (UID: "5da5ddac-6e96-4800-b178-a03b3cdc1dfa"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.757625 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.758192 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:03:40 crc kubenswrapper[4956]: I1126 17:03:40.758209 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nt4wv\" (UniqueName: \"kubernetes.io/projected/5da5ddac-6e96-4800-b178-a03b3cdc1dfa-kube-api-access-nt4wv\") on node \"crc\" DevicePath \"\"" Nov 26 17:03:41 crc kubenswrapper[4956]: I1126 17:03:41.277288 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" event={"ID":"5da5ddac-6e96-4800-b178-a03b3cdc1dfa","Type":"ContainerDied","Data":"1c024cb7b74bbdc7fb72b41f93b8859e6a2fd50ce4cfbf65741fe8b2855d626b"} Nov 26 17:03:41 crc kubenswrapper[4956]: I1126 17:03:41.277347 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv" Nov 26 17:03:41 crc kubenswrapper[4956]: I1126 17:03:41.277377 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1c024cb7b74bbdc7fb72b41f93b8859e6a2fd50ce4cfbf65741fe8b2855d626b" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.740579 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4"] Nov 26 17:03:53 crc kubenswrapper[4956]: E1126 17:03:53.741491 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="util" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.741506 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="util" Nov 26 17:03:53 crc kubenswrapper[4956]: E1126 17:03:53.741516 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="pull" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.741522 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="pull" Nov 26 17:03:53 crc kubenswrapper[4956]: E1126 17:03:53.741539 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="extract" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.741545 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="extract" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.741645 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5da5ddac-6e96-4800-b178-a03b3cdc1dfa" containerName="extract" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.742102 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.745723 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.745910 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.745923 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.746277 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rwdd4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.747192 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.796899 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4"] Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.894723 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-apiservice-cert\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.894934 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dlzf2\" (UniqueName: \"kubernetes.io/projected/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-kube-api-access-dlzf2\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.895019 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-webhook-cert\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.995914 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dlzf2\" (UniqueName: \"kubernetes.io/projected/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-kube-api-access-dlzf2\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.995984 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-webhook-cert\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.996035 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-apiservice-cert\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.996382 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk"] Nov 26 17:03:53 crc kubenswrapper[4956]: I1126 17:03:53.997206 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.000288 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.000783 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-t7s7x" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.001021 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.004470 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-apiservice-cert\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.006597 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-webhook-cert\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.006762 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk"] Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.021616 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dlzf2\" (UniqueName: \"kubernetes.io/projected/4b5ac5a5-46f2-4fc2-944d-58fcafb997b5-kube-api-access-dlzf2\") pod \"metallb-operator-controller-manager-6d556748c5-l2bl4\" (UID: \"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5\") " pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.074930 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.097716 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jv4r4\" (UniqueName: \"kubernetes.io/projected/d7584f30-fba3-40ac-b6f2-881664bc6d6a-kube-api-access-jv4r4\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.097926 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7584f30-fba3-40ac-b6f2-881664bc6d6a-apiservice-cert\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.098013 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7584f30-fba3-40ac-b6f2-881664bc6d6a-webhook-cert\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.199362 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7584f30-fba3-40ac-b6f2-881664bc6d6a-apiservice-cert\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.199850 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7584f30-fba3-40ac-b6f2-881664bc6d6a-webhook-cert\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.199916 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jv4r4\" (UniqueName: \"kubernetes.io/projected/d7584f30-fba3-40ac-b6f2-881664bc6d6a-kube-api-access-jv4r4\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.207070 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d7584f30-fba3-40ac-b6f2-881664bc6d6a-apiservice-cert\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.217531 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d7584f30-fba3-40ac-b6f2-881664bc6d6a-webhook-cert\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.217762 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jv4r4\" (UniqueName: \"kubernetes.io/projected/d7584f30-fba3-40ac-b6f2-881664bc6d6a-kube-api-access-jv4r4\") pod \"metallb-operator-webhook-server-7bcbfbd867-h25xk\" (UID: \"d7584f30-fba3-40ac-b6f2-881664bc6d6a\") " pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.358627 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4"] Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.358758 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:03:54 crc kubenswrapper[4956]: I1126 17:03:54.591759 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk"] Nov 26 17:03:54 crc kubenswrapper[4956]: W1126 17:03:54.593206 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd7584f30_fba3_40ac_b6f2_881664bc6d6a.slice/crio-a62c519fae7e714a2ffb6af53d696f87b3881136026a48967eb5ac1edc8d85b0 WatchSource:0}: Error finding container a62c519fae7e714a2ffb6af53d696f87b3881136026a48967eb5ac1edc8d85b0: Status 404 returned error can't find the container with id a62c519fae7e714a2ffb6af53d696f87b3881136026a48967eb5ac1edc8d85b0 Nov 26 17:03:55 crc kubenswrapper[4956]: I1126 17:03:55.382825 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" event={"ID":"d7584f30-fba3-40ac-b6f2-881664bc6d6a","Type":"ContainerStarted","Data":"a62c519fae7e714a2ffb6af53d696f87b3881136026a48967eb5ac1edc8d85b0"} Nov 26 17:03:55 crc kubenswrapper[4956]: I1126 17:03:55.385125 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" event={"ID":"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5","Type":"ContainerStarted","Data":"c3d9d0a64862607f85f1a462b80739278dad391326d1b528edb5b3cbb3bd2b42"} Nov 26 17:04:01 crc kubenswrapper[4956]: I1126 17:04:01.422538 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" event={"ID":"d7584f30-fba3-40ac-b6f2-881664bc6d6a","Type":"ContainerStarted","Data":"5872bb1b728aac6f739895ca9b13b55ec429ffd9775dbb0281a3db15aedfc9bb"} Nov 26 17:04:01 crc kubenswrapper[4956]: I1126 17:04:01.424103 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:04:01 crc kubenswrapper[4956]: I1126 17:04:01.425967 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" event={"ID":"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5","Type":"ContainerStarted","Data":"c9a78b3017292779ce3a1dab640bfee03067a0aa57e9353b9942d952daa07a7a"} Nov 26 17:04:01 crc kubenswrapper[4956]: I1126 17:04:01.426546 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:04:01 crc kubenswrapper[4956]: I1126 17:04:01.467805 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" podStartSLOduration=2.583326377 podStartE2EDuration="8.467777553s" podCreationTimestamp="2025-11-26 17:03:53 +0000 UTC" firstStartedPulling="2025-11-26 17:03:54.371966433 +0000 UTC m=+680.067926985" lastFinishedPulling="2025-11-26 17:04:00.256417609 +0000 UTC m=+685.952378161" observedRunningTime="2025-11-26 17:04:01.463855462 +0000 UTC m=+687.159816024" watchObservedRunningTime="2025-11-26 17:04:01.467777553 +0000 UTC m=+687.163738105" Nov 26 17:04:01 crc kubenswrapper[4956]: I1126 17:04:01.468078 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" podStartSLOduration=2.790542872 podStartE2EDuration="8.468072162s" podCreationTimestamp="2025-11-26 17:03:53 +0000 UTC" firstStartedPulling="2025-11-26 17:03:54.597770616 +0000 UTC m=+680.293731168" lastFinishedPulling="2025-11-26 17:04:00.275299906 +0000 UTC m=+685.971260458" observedRunningTime="2025-11-26 17:04:01.444885342 +0000 UTC m=+687.140845894" watchObservedRunningTime="2025-11-26 17:04:01.468072162 +0000 UTC m=+687.164032714" Nov 26 17:04:14 crc kubenswrapper[4956]: I1126 17:04:14.365683 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-7bcbfbd867-h25xk" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.081685 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.771830 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-79tsd"] Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.775173 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.777499 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.777653 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-rr2js" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.777980 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.789062 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h"] Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.789900 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.792392 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.803049 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h"] Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863347 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-metrics\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863419 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-frr-sockets\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863469 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2ca64dff-1360-4261-81c8-61192f37b99a-frr-startup\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863503 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vld89\" (UniqueName: \"kubernetes.io/projected/2ca64dff-1360-4261-81c8-61192f37b99a-kube-api-access-vld89\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863549 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-frr-conf\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863567 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-reloader\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.863604 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ca64dff-1360-4261-81c8-61192f37b99a-metrics-certs\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.892804 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-768ll"] Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.894314 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-768ll" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.895430 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-6c7b4b5f48-rwzwq"] Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.896475 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.897663 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.897932 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.898064 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-ldx5h" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.898185 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.900901 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.922205 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-rwzwq"] Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964538 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2ca64dff-1360-4261-81c8-61192f37b99a-frr-startup\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964603 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-46gkg\" (UniqueName: \"kubernetes.io/projected/73efb92c-5732-443b-8f6f-a1b183f7a455-kube-api-access-46gkg\") pod \"frr-k8s-webhook-server-6998585d5-wnm7h\" (UID: \"73efb92c-5732-443b-8f6f-a1b183f7a455\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964629 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vld89\" (UniqueName: \"kubernetes.io/projected/2ca64dff-1360-4261-81c8-61192f37b99a-kube-api-access-vld89\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964649 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-frr-conf\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964665 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-reloader\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964684 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73efb92c-5732-443b-8f6f-a1b183f7a455-cert\") pod \"frr-k8s-webhook-server-6998585d5-wnm7h\" (UID: \"73efb92c-5732-443b-8f6f-a1b183f7a455\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964717 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ca64dff-1360-4261-81c8-61192f37b99a-metrics-certs\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964742 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-metrics\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.964757 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-frr-sockets\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.965560 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-frr-sockets\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.965743 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-reloader\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.965991 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-metrics\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.966373 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/2ca64dff-1360-4261-81c8-61192f37b99a-frr-conf\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.967278 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.967769 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.976068 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/2ca64dff-1360-4261-81c8-61192f37b99a-frr-startup\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:34 crc kubenswrapper[4956]: I1126 17:04:34.996678 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/2ca64dff-1360-4261-81c8-61192f37b99a-metrics-certs\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.022034 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vld89\" (UniqueName: \"kubernetes.io/projected/2ca64dff-1360-4261-81c8-61192f37b99a-kube-api-access-vld89\") pod \"frr-k8s-79tsd\" (UID: \"2ca64dff-1360-4261-81c8-61192f37b99a\") " pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066301 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-memberlist\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066380 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-46gkg\" (UniqueName: \"kubernetes.io/projected/73efb92c-5732-443b-8f6f-a1b183f7a455-kube-api-access-46gkg\") pod \"frr-k8s-webhook-server-6998585d5-wnm7h\" (UID: \"73efb92c-5732-443b-8f6f-a1b183f7a455\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066427 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z65ln\" (UniqueName: \"kubernetes.io/projected/1891d982-bc1a-4807-b53f-cf1837eb0c0b-kube-api-access-z65ln\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066459 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73efb92c-5732-443b-8f6f-a1b183f7a455-cert\") pod \"frr-k8s-webhook-server-6998585d5-wnm7h\" (UID: \"73efb92c-5732-443b-8f6f-a1b183f7a455\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066679 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1891d982-bc1a-4807-b53f-cf1837eb0c0b-metrics-certs\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066782 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/86b7d34e-df87-4f72-a20b-d8f241516ac9-metallb-excludel2\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.066801 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1891d982-bc1a-4807-b53f-cf1837eb0c0b-cert\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.068075 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-metrics-certs\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.068129 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvhj5\" (UniqueName: \"kubernetes.io/projected/86b7d34e-df87-4f72-a20b-d8f241516ac9-kube-api-access-bvhj5\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.069211 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.081651 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/73efb92c-5732-443b-8f6f-a1b183f7a455-cert\") pod \"frr-k8s-webhook-server-6998585d5-wnm7h\" (UID: \"73efb92c-5732-443b-8f6f-a1b183f7a455\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.083616 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-46gkg\" (UniqueName: \"kubernetes.io/projected/73efb92c-5732-443b-8f6f-a1b183f7a455-kube-api-access-46gkg\") pod \"frr-k8s-webhook-server-6998585d5-wnm7h\" (UID: \"73efb92c-5732-443b-8f6f-a1b183f7a455\") " pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.104389 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-rr2js" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.111177 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.114912 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.172819 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/86b7d34e-df87-4f72-a20b-d8f241516ac9-metallb-excludel2\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.172901 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1891d982-bc1a-4807-b53f-cf1837eb0c0b-cert\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.172957 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-metrics-certs\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.172985 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvhj5\" (UniqueName: \"kubernetes.io/projected/86b7d34e-df87-4f72-a20b-d8f241516ac9-kube-api-access-bvhj5\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.173017 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-memberlist\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.173052 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z65ln\" (UniqueName: \"kubernetes.io/projected/1891d982-bc1a-4807-b53f-cf1837eb0c0b-kube-api-access-z65ln\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.173131 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1891d982-bc1a-4807-b53f-cf1837eb0c0b-metrics-certs\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.179815 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.180032 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.180212 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.180369 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.180479 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 26 17:04:35 crc kubenswrapper[4956]: E1126 17:04:35.185028 4956 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Nov 26 17:04:35 crc kubenswrapper[4956]: E1126 17:04:35.185114 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-memberlist podName:86b7d34e-df87-4f72-a20b-d8f241516ac9 nodeName:}" failed. No retries permitted until 2025-11-26 17:04:35.68508718 +0000 UTC m=+721.381047742 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-memberlist") pod "speaker-768ll" (UID: "86b7d34e-df87-4f72-a20b-d8f241516ac9") : secret "metallb-memberlist" not found Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.186098 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/86b7d34e-df87-4f72-a20b-d8f241516ac9-metallb-excludel2\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.191183 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/1891d982-bc1a-4807-b53f-cf1837eb0c0b-metrics-certs\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.192747 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1891d982-bc1a-4807-b53f-cf1837eb0c0b-cert\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.193402 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-metrics-certs\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.206786 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z65ln\" (UniqueName: \"kubernetes.io/projected/1891d982-bc1a-4807-b53f-cf1837eb0c0b-kube-api-access-z65ln\") pod \"controller-6c7b4b5f48-rwzwq\" (UID: \"1891d982-bc1a-4807-b53f-cf1837eb0c0b\") " pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.210429 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvhj5\" (UniqueName: \"kubernetes.io/projected/86b7d34e-df87-4f72-a20b-d8f241516ac9-kube-api-access-bvhj5\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.219688 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.407697 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h"] Nov 26 17:04:35 crc kubenswrapper[4956]: W1126 17:04:35.414729 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod73efb92c_5732_443b_8f6f_a1b183f7a455.slice/crio-162463af493c013f534a75d3135bb345dfd2db958d483338f483455fb8dfded2 WatchSource:0}: Error finding container 162463af493c013f534a75d3135bb345dfd2db958d483338f483455fb8dfded2: Status 404 returned error can't find the container with id 162463af493c013f534a75d3135bb345dfd2db958d483338f483455fb8dfded2 Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.470410 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-6c7b4b5f48-rwzwq"] Nov 26 17:04:35 crc kubenswrapper[4956]: W1126 17:04:35.478268 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1891d982_bc1a_4807_b53f_cf1837eb0c0b.slice/crio-9e3eb19f6383c8ccd0c78f32441f362edc512ecacbafb484242ace0dbe023ea5 WatchSource:0}: Error finding container 9e3eb19f6383c8ccd0c78f32441f362edc512ecacbafb484242ace0dbe023ea5: Status 404 returned error can't find the container with id 9e3eb19f6383c8ccd0c78f32441f362edc512ecacbafb484242ace0dbe023ea5 Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.641000 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-rwzwq" event={"ID":"1891d982-bc1a-4807-b53f-cf1837eb0c0b","Type":"ContainerStarted","Data":"e3a3c44ef89c1ec687d7e7d2001a9db288535175f0bf73db5bf0ce17bb572fb4"} Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.641062 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-rwzwq" event={"ID":"1891d982-bc1a-4807-b53f-cf1837eb0c0b","Type":"ContainerStarted","Data":"9e3eb19f6383c8ccd0c78f32441f362edc512ecacbafb484242ace0dbe023ea5"} Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.642554 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" event={"ID":"73efb92c-5732-443b-8f6f-a1b183f7a455","Type":"ContainerStarted","Data":"162463af493c013f534a75d3135bb345dfd2db958d483338f483455fb8dfded2"} Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.643545 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"050b66c660165d68e3e5d0f0d31efb20ae9b4725ab740bcdc6c36810f3deb533"} Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.781950 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-memberlist\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.791651 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/86b7d34e-df87-4f72-a20b-d8f241516ac9-memberlist\") pod \"speaker-768ll\" (UID: \"86b7d34e-df87-4f72-a20b-d8f241516ac9\") " pod="metallb-system/speaker-768ll" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.815888 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-ldx5h" Nov 26 17:04:35 crc kubenswrapper[4956]: I1126 17:04:35.822924 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-768ll" Nov 26 17:04:36 crc kubenswrapper[4956]: I1126 17:04:36.653457 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-768ll" event={"ID":"86b7d34e-df87-4f72-a20b-d8f241516ac9","Type":"ContainerStarted","Data":"4155e0be1a72d80fc79dd94abeb9002dc7ad6b99cdea23eef1aac2fc984ee2b2"} Nov 26 17:04:36 crc kubenswrapper[4956]: I1126 17:04:36.653509 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-768ll" event={"ID":"86b7d34e-df87-4f72-a20b-d8f241516ac9","Type":"ContainerStarted","Data":"05fe2d592a9606da00b6fee0d0add1b4a528e59370b752eee305950dfb8c8fcb"} Nov 26 17:04:39 crc kubenswrapper[4956]: I1126 17:04:39.551461 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:04:39 crc kubenswrapper[4956]: I1126 17:04:39.552163 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.733054 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-768ll" event={"ID":"86b7d34e-df87-4f72-a20b-d8f241516ac9","Type":"ContainerStarted","Data":"ca670c75571eb6ba641d2993ebb33bed581541ba9211386f3a858d828451e7f0"} Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.733926 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-768ll" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.735015 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-6c7b4b5f48-rwzwq" event={"ID":"1891d982-bc1a-4807-b53f-cf1837eb0c0b","Type":"ContainerStarted","Data":"09beec21018048443ba232086434615078996d4de7967ad52e4f324dca07f645"} Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.735183 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.737188 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" event={"ID":"73efb92c-5732-443b-8f6f-a1b183f7a455","Type":"ContainerStarted","Data":"a93ab28ff351bb431a8f2ec9c02b6bf6e2bdab0c933b78c9fc955b32cc08fce9"} Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.737611 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.739854 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-768ll" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.742575 4956 generic.go:334] "Generic (PLEG): container finished" podID="2ca64dff-1360-4261-81c8-61192f37b99a" containerID="52aaef97b08b6c0044a1ec034e8e1239b50618d0f2700bb647e902780cff67c6" exitCode=0 Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.742623 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerDied","Data":"52aaef97b08b6c0044a1ec034e8e1239b50618d0f2700bb647e902780cff67c6"} Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.743441 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-6c7b4b5f48-rwzwq" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.763126 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-768ll" podStartSLOduration=3.147492577 podStartE2EDuration="12.763105647s" podCreationTimestamp="2025-11-26 17:04:34 +0000 UTC" firstStartedPulling="2025-11-26 17:04:36.251026609 +0000 UTC m=+721.946987161" lastFinishedPulling="2025-11-26 17:04:45.866639679 +0000 UTC m=+731.562600231" observedRunningTime="2025-11-26 17:04:46.760540073 +0000 UTC m=+732.456500635" watchObservedRunningTime="2025-11-26 17:04:46.763105647 +0000 UTC m=+732.459066199" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.785530 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" podStartSLOduration=1.841579723 podStartE2EDuration="12.785508847s" podCreationTimestamp="2025-11-26 17:04:34 +0000 UTC" firstStartedPulling="2025-11-26 17:04:35.418306742 +0000 UTC m=+721.114267284" lastFinishedPulling="2025-11-26 17:04:46.362235856 +0000 UTC m=+732.058196408" observedRunningTime="2025-11-26 17:04:46.781824861 +0000 UTC m=+732.477785413" watchObservedRunningTime="2025-11-26 17:04:46.785508847 +0000 UTC m=+732.481469399" Nov 26 17:04:46 crc kubenswrapper[4956]: I1126 17:04:46.866532 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-6c7b4b5f48-rwzwq" podStartSLOduration=2.098386637 podStartE2EDuration="12.86651467s" podCreationTimestamp="2025-11-26 17:04:34 +0000 UTC" firstStartedPulling="2025-11-26 17:04:35.64650773 +0000 UTC m=+721.342468282" lastFinishedPulling="2025-11-26 17:04:46.414635753 +0000 UTC m=+732.110596315" observedRunningTime="2025-11-26 17:04:46.833087505 +0000 UTC m=+732.529048067" watchObservedRunningTime="2025-11-26 17:04:46.86651467 +0000 UTC m=+732.562475222" Nov 26 17:04:47 crc kubenswrapper[4956]: I1126 17:04:47.751056 4956 generic.go:334] "Generic (PLEG): container finished" podID="2ca64dff-1360-4261-81c8-61192f37b99a" containerID="db3e320fdc2c60fdab17252c03cae38b5ea99eadf481149cd0d8279ad7b40f04" exitCode=0 Nov 26 17:04:47 crc kubenswrapper[4956]: I1126 17:04:47.751232 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerDied","Data":"db3e320fdc2c60fdab17252c03cae38b5ea99eadf481149cd0d8279ad7b40f04"} Nov 26 17:04:48 crc kubenswrapper[4956]: I1126 17:04:48.767627 4956 generic.go:334] "Generic (PLEG): container finished" podID="2ca64dff-1360-4261-81c8-61192f37b99a" containerID="da208fc863daa74b664c7fb08dc5c27cd10892bc6744e088ece6ced4b494beee" exitCode=0 Nov 26 17:04:48 crc kubenswrapper[4956]: I1126 17:04:48.767769 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerDied","Data":"da208fc863daa74b664c7fb08dc5c27cd10892bc6744e088ece6ced4b494beee"} Nov 26 17:04:49 crc kubenswrapper[4956]: I1126 17:04:49.777568 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"7da1da948984a21c8f18fbcb84c06c541967047036e242c3c68e087ef921c9a9"} Nov 26 17:04:49 crc kubenswrapper[4956]: I1126 17:04:49.777976 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"85c2627ec61121ea5cb2c7072f08f9ceb8bca5bfc1401315194c2e948ef38b80"} Nov 26 17:04:50 crc kubenswrapper[4956]: I1126 17:04:50.794833 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"a69914857f04134e8a4150691904c1fdd17c1ec270add57e6db138db034ffd89"} Nov 26 17:04:50 crc kubenswrapper[4956]: I1126 17:04:50.795678 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"0c012bd57696933e352d195102b41619b65a92ddc7de87e4f57a12a14ff87fc6"} Nov 26 17:04:50 crc kubenswrapper[4956]: I1126 17:04:50.796087 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"9781b8a3fd180d030becdc1bb1842c01fd1ce930fbc7b9e5af6e5aa4c74f6f64"} Nov 26 17:04:51 crc kubenswrapper[4956]: I1126 17:04:51.811115 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-79tsd" event={"ID":"2ca64dff-1360-4261-81c8-61192f37b99a","Type":"ContainerStarted","Data":"2f948767efe15be3ff95928973c8548d6d8010104026a9423fa886de68a2ec3f"} Nov 26 17:04:51 crc kubenswrapper[4956]: I1126 17:04:51.812337 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:51 crc kubenswrapper[4956]: I1126 17:04:51.848845 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-79tsd" podStartSLOduration=6.829683205 podStartE2EDuration="17.848824628s" podCreationTimestamp="2025-11-26 17:04:34 +0000 UTC" firstStartedPulling="2025-11-26 17:04:35.343200336 +0000 UTC m=+721.039160888" lastFinishedPulling="2025-11-26 17:04:46.362341759 +0000 UTC m=+732.058302311" observedRunningTime="2025-11-26 17:04:51.843777454 +0000 UTC m=+737.539738026" watchObservedRunningTime="2025-11-26 17:04:51.848824628 +0000 UTC m=+737.544785180" Nov 26 17:04:53 crc kubenswrapper[4956]: I1126 17:04:53.742959 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-d92b2"] Nov 26 17:04:53 crc kubenswrapper[4956]: I1126 17:04:53.744462 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:04:53 crc kubenswrapper[4956]: W1126 17:04:53.746642 4956 reflector.go:561] object-"openstack-operators"/"openshift-service-ca.crt": failed to list *v1.ConfigMap: configmaps "openshift-service-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Nov 26 17:04:53 crc kubenswrapper[4956]: E1126 17:04:53.746711 4956 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"openshift-service-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"openshift-service-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 26 17:04:53 crc kubenswrapper[4956]: W1126 17:04:53.748485 4956 reflector.go:561] object-"openstack-operators"/"mariadb-operator-index-dockercfg-l8tt5": failed to list *v1.Secret: secrets "mariadb-operator-index-dockercfg-l8tt5" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Nov 26 17:04:53 crc kubenswrapper[4956]: E1126 17:04:53.748657 4956 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"mariadb-operator-index-dockercfg-l8tt5\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"mariadb-operator-index-dockercfg-l8tt5\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 26 17:04:53 crc kubenswrapper[4956]: W1126 17:04:53.748532 4956 reflector.go:561] object-"openstack-operators"/"kube-root-ca.crt": failed to list *v1.ConfigMap: configmaps "kube-root-ca.crt" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack-operators": no relationship found between node 'crc' and this object Nov 26 17:04:53 crc kubenswrapper[4956]: E1126 17:04:53.748842 4956 reflector.go:158] "Unhandled Error" err="object-\"openstack-operators\"/\"kube-root-ca.crt\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"kube-root-ca.crt\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack-operators\": no relationship found between node 'crc' and this object" logger="UnhandledError" Nov 26 17:04:53 crc kubenswrapper[4956]: I1126 17:04:53.750065 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-d92b2"] Nov 26 17:04:53 crc kubenswrapper[4956]: I1126 17:04:53.831228 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwkdb\" (UniqueName: \"kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb\") pod \"mariadb-operator-index-d92b2\" (UID: \"21d7cda2-cd2b-4ad8-911f-e117e129381f\") " pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:04:53 crc kubenswrapper[4956]: I1126 17:04:53.932973 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkdb\" (UniqueName: \"kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb\") pod \"mariadb-operator-index-d92b2\" (UID: \"21d7cda2-cd2b-4ad8-911f-e117e129381f\") " pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:04:54 crc kubenswrapper[4956]: E1126 17:04:54.947287 4956 projected.go:288] Couldn't get configMap openstack-operators/kube-root-ca.crt: failed to sync configmap cache: timed out waiting for the condition Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.115660 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.176988 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-79tsd" Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.212502 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-l8tt5" Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.271748 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.303276 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 26 17:04:55 crc kubenswrapper[4956]: E1126 17:04:55.308453 4956 projected.go:194] Error preparing data for projected volume kube-api-access-mwkdb for pod openstack-operators/mariadb-operator-index-d92b2: failed to sync configmap cache: timed out waiting for the condition Nov 26 17:04:55 crc kubenswrapper[4956]: E1126 17:04:55.308580 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb podName:21d7cda2-cd2b-4ad8-911f-e117e129381f nodeName:}" failed. No retries permitted until 2025-11-26 17:04:55.808546165 +0000 UTC m=+741.504506747 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-mwkdb" (UniqueName: "kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb") pod "mariadb-operator-index-d92b2" (UID: "21d7cda2-cd2b-4ad8-911f-e117e129381f") : failed to sync configmap cache: timed out waiting for the condition Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.861808 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwkdb\" (UniqueName: \"kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb\") pod \"mariadb-operator-index-d92b2\" (UID: \"21d7cda2-cd2b-4ad8-911f-e117e129381f\") " pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:04:55 crc kubenswrapper[4956]: I1126 17:04:55.872716 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwkdb\" (UniqueName: \"kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb\") pod \"mariadb-operator-index-d92b2\" (UID: \"21d7cda2-cd2b-4ad8-911f-e117e129381f\") " pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:04:56 crc kubenswrapper[4956]: I1126 17:04:56.170534 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:04:56 crc kubenswrapper[4956]: I1126 17:04:56.632463 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-d92b2"] Nov 26 17:04:56 crc kubenswrapper[4956]: W1126 17:04:56.646069 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod21d7cda2_cd2b_4ad8_911f_e117e129381f.slice/crio-6f708d248ddbae05fca5931d2d7ee7aacfa700adcdcc7b44df01ef7ccb3fbc5d WatchSource:0}: Error finding container 6f708d248ddbae05fca5931d2d7ee7aacfa700adcdcc7b44df01ef7ccb3fbc5d: Status 404 returned error can't find the container with id 6f708d248ddbae05fca5931d2d7ee7aacfa700adcdcc7b44df01ef7ccb3fbc5d Nov 26 17:04:56 crc kubenswrapper[4956]: I1126 17:04:56.850129 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d92b2" event={"ID":"21d7cda2-cd2b-4ad8-911f-e117e129381f","Type":"ContainerStarted","Data":"6f708d248ddbae05fca5931d2d7ee7aacfa700adcdcc7b44df01ef7ccb3fbc5d"} Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.118361 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-d92b2"] Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.723816 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-index-w2wwv"] Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.724787 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.765963 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-w2wwv"] Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.790924 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g4jlz\" (UniqueName: \"kubernetes.io/projected/cd46bd02-d445-4a8b-8712-b86b9ca6c0d2-kube-api-access-g4jlz\") pod \"mariadb-operator-index-w2wwv\" (UID: \"cd46bd02-d445-4a8b-8712-b86b9ca6c0d2\") " pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.892547 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g4jlz\" (UniqueName: \"kubernetes.io/projected/cd46bd02-d445-4a8b-8712-b86b9ca6c0d2-kube-api-access-g4jlz\") pod \"mariadb-operator-index-w2wwv\" (UID: \"cd46bd02-d445-4a8b-8712-b86b9ca6c0d2\") " pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:04:57 crc kubenswrapper[4956]: I1126 17:04:57.915950 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g4jlz\" (UniqueName: \"kubernetes.io/projected/cd46bd02-d445-4a8b-8712-b86b9ca6c0d2-kube-api-access-g4jlz\") pod \"mariadb-operator-index-w2wwv\" (UID: \"cd46bd02-d445-4a8b-8712-b86b9ca6c0d2\") " pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:04:58 crc kubenswrapper[4956]: I1126 17:04:58.081711 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:04:58 crc kubenswrapper[4956]: I1126 17:04:58.296247 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-index-w2wwv"] Nov 26 17:04:58 crc kubenswrapper[4956]: W1126 17:04:58.299705 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd46bd02_d445_4a8b_8712_b86b9ca6c0d2.slice/crio-9713e58c43f56d7c42a19cc933b672530fb2d01ba8100e96bdbc14355e494319 WatchSource:0}: Error finding container 9713e58c43f56d7c42a19cc933b672530fb2d01ba8100e96bdbc14355e494319: Status 404 returned error can't find the container with id 9713e58c43f56d7c42a19cc933b672530fb2d01ba8100e96bdbc14355e494319 Nov 26 17:04:58 crc kubenswrapper[4956]: I1126 17:04:58.866375 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-w2wwv" event={"ID":"cd46bd02-d445-4a8b-8712-b86b9ca6c0d2","Type":"ContainerStarted","Data":"9713e58c43f56d7c42a19cc933b672530fb2d01ba8100e96bdbc14355e494319"} Nov 26 17:05:00 crc kubenswrapper[4956]: I1126 17:05:00.883556 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d92b2" event={"ID":"21d7cda2-cd2b-4ad8-911f-e117e129381f","Type":"ContainerStarted","Data":"6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46"} Nov 26 17:05:00 crc kubenswrapper[4956]: I1126 17:05:00.883803 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/mariadb-operator-index-d92b2" podUID="21d7cda2-cd2b-4ad8-911f-e117e129381f" containerName="registry-server" containerID="cri-o://6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46" gracePeriod=2 Nov 26 17:05:00 crc kubenswrapper[4956]: I1126 17:05:00.885224 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-w2wwv" event={"ID":"cd46bd02-d445-4a8b-8712-b86b9ca6c0d2","Type":"ContainerStarted","Data":"9f38065d1a32141390ee2bff6d1d2c687c99341cc9d5bb1cddeaed52c6047a2d"} Nov 26 17:05:00 crc kubenswrapper[4956]: I1126 17:05:00.907636 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-d92b2" podStartSLOduration=4.752187346 podStartE2EDuration="7.907611991s" podCreationTimestamp="2025-11-26 17:04:53 +0000 UTC" firstStartedPulling="2025-11-26 17:04:56.6502463 +0000 UTC m=+742.346206882" lastFinishedPulling="2025-11-26 17:04:59.805670975 +0000 UTC m=+745.501631527" observedRunningTime="2025-11-26 17:05:00.900741075 +0000 UTC m=+746.596701647" watchObservedRunningTime="2025-11-26 17:05:00.907611991 +0000 UTC m=+746.603572543" Nov 26 17:05:00 crc kubenswrapper[4956]: I1126 17:05:00.919570 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-index-w2wwv" podStartSLOduration=2.41799952 podStartE2EDuration="3.919545752s" podCreationTimestamp="2025-11-26 17:04:57 +0000 UTC" firstStartedPulling="2025-11-26 17:04:58.302807806 +0000 UTC m=+743.998768358" lastFinishedPulling="2025-11-26 17:04:59.804354028 +0000 UTC m=+745.500314590" observedRunningTime="2025-11-26 17:05:00.918744529 +0000 UTC m=+746.614705081" watchObservedRunningTime="2025-11-26 17:05:00.919545752 +0000 UTC m=+746.615506314" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.248833 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.358809 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mwkdb\" (UniqueName: \"kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb\") pod \"21d7cda2-cd2b-4ad8-911f-e117e129381f\" (UID: \"21d7cda2-cd2b-4ad8-911f-e117e129381f\") " Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.366253 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb" (OuterVolumeSpecName: "kube-api-access-mwkdb") pod "21d7cda2-cd2b-4ad8-911f-e117e129381f" (UID: "21d7cda2-cd2b-4ad8-911f-e117e129381f"). InnerVolumeSpecName "kube-api-access-mwkdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.461058 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mwkdb\" (UniqueName: \"kubernetes.io/projected/21d7cda2-cd2b-4ad8-911f-e117e129381f-kube-api-access-mwkdb\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.899994 4956 generic.go:334] "Generic (PLEG): container finished" podID="21d7cda2-cd2b-4ad8-911f-e117e129381f" containerID="6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46" exitCode=0 Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.901229 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-index-d92b2" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.902583 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d92b2" event={"ID":"21d7cda2-cd2b-4ad8-911f-e117e129381f","Type":"ContainerDied","Data":"6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46"} Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.902695 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-index-d92b2" event={"ID":"21d7cda2-cd2b-4ad8-911f-e117e129381f","Type":"ContainerDied","Data":"6f708d248ddbae05fca5931d2d7ee7aacfa700adcdcc7b44df01ef7ccb3fbc5d"} Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.902923 4956 scope.go:117] "RemoveContainer" containerID="6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.920397 4956 scope.go:117] "RemoveContainer" containerID="6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46" Nov 26 17:05:01 crc kubenswrapper[4956]: E1126 17:05:01.921144 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46\": container with ID starting with 6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46 not found: ID does not exist" containerID="6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.921244 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46"} err="failed to get container status \"6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46\": rpc error: code = NotFound desc = could not find container \"6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46\": container with ID starting with 6c8d6e8b3f8b0e29007bbc6ae54a2467c116904ea07ad66b56725001c0e05d46 not found: ID does not exist" Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.942093 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/mariadb-operator-index-d92b2"] Nov 26 17:05:01 crc kubenswrapper[4956]: I1126 17:05:01.946902 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/mariadb-operator-index-d92b2"] Nov 26 17:05:03 crc kubenswrapper[4956]: I1126 17:05:03.003957 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="21d7cda2-cd2b-4ad8-911f-e117e129381f" path="/var/lib/kubelet/pods/21d7cda2-cd2b-4ad8-911f-e117e129381f/volumes" Nov 26 17:05:05 crc kubenswrapper[4956]: I1126 17:05:05.118403 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-6998585d5-wnm7h" Nov 26 17:05:05 crc kubenswrapper[4956]: I1126 17:05:05.120022 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-79tsd" Nov 26 17:05:08 crc kubenswrapper[4956]: I1126 17:05:08.083053 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:05:08 crc kubenswrapper[4956]: I1126 17:05:08.083742 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:05:08 crc kubenswrapper[4956]: I1126 17:05:08.122993 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:05:08 crc kubenswrapper[4956]: I1126 17:05:08.991752 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-index-w2wwv" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.338407 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2"] Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.338858 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerName="route-controller-manager" containerID="cri-o://932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda" gracePeriod=30 Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.348614 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v5xlx"] Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.349517 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" podUID="b9b65044-9326-4a72-a933-e84012a29211" containerName="controller-manager" containerID="cri-o://7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753" gracePeriod=30 Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.550885 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.550973 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.900402 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.907357 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.963108 4956 generic.go:334] "Generic (PLEG): container finished" podID="b9b65044-9326-4a72-a933-e84012a29211" containerID="7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753" exitCode=0 Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.963176 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" event={"ID":"b9b65044-9326-4a72-a933-e84012a29211","Type":"ContainerDied","Data":"7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753"} Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.963215 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" event={"ID":"b9b65044-9326-4a72-a933-e84012a29211","Type":"ContainerDied","Data":"14fc8934430a0279fc8fe39e01d9feaa7897869a5cbc4187b006ec032291d4f8"} Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.963235 4956 scope.go:117] "RemoveContainer" containerID="7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.963350 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-v5xlx" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.967403 4956 generic.go:334] "Generic (PLEG): container finished" podID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerID="932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda" exitCode=0 Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.967465 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.967517 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" event={"ID":"fa92cf8a-56c1-4b38-889c-23a578b74b90","Type":"ContainerDied","Data":"932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda"} Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.967581 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2" event={"ID":"fa92cf8a-56c1-4b38-889c-23a578b74b90","Type":"ContainerDied","Data":"e9dadad485fad8cfb5c1f50a10dc8db68a6177400d6b2892e650b410d69d0de7"} Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.987634 4956 scope.go:117] "RemoveContainer" containerID="7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753" Nov 26 17:05:09 crc kubenswrapper[4956]: E1126 17:05:09.988622 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753\": container with ID starting with 7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753 not found: ID does not exist" containerID="7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.988679 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753"} err="failed to get container status \"7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753\": rpc error: code = NotFound desc = could not find container \"7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753\": container with ID starting with 7e0dd2c45944ed69a8384b24c2227a7f87c507b3599cfb00f814ac5c55a59753 not found: ID does not exist" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.988711 4956 scope.go:117] "RemoveContainer" containerID="932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989088 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-config\") pod \"b9b65044-9326-4a72-a933-e84012a29211\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989177 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-config\") pod \"fa92cf8a-56c1-4b38-889c-23a578b74b90\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989216 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b65044-9326-4a72-a933-e84012a29211-serving-cert\") pod \"b9b65044-9326-4a72-a933-e84012a29211\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989257 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa92cf8a-56c1-4b38-889c-23a578b74b90-serving-cert\") pod \"fa92cf8a-56c1-4b38-889c-23a578b74b90\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989306 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-client-ca\") pod \"b9b65044-9326-4a72-a933-e84012a29211\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989333 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-client-ca\") pod \"fa92cf8a-56c1-4b38-889c-23a578b74b90\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989365 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q57kt\" (UniqueName: \"kubernetes.io/projected/fa92cf8a-56c1-4b38-889c-23a578b74b90-kube-api-access-q57kt\") pod \"fa92cf8a-56c1-4b38-889c-23a578b74b90\" (UID: \"fa92cf8a-56c1-4b38-889c-23a578b74b90\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989422 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vxsrf\" (UniqueName: \"kubernetes.io/projected/b9b65044-9326-4a72-a933-e84012a29211-kube-api-access-vxsrf\") pod \"b9b65044-9326-4a72-a933-e84012a29211\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.989490 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-proxy-ca-bundles\") pod \"b9b65044-9326-4a72-a933-e84012a29211\" (UID: \"b9b65044-9326-4a72-a933-e84012a29211\") " Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.990673 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "b9b65044-9326-4a72-a933-e84012a29211" (UID: "b9b65044-9326-4a72-a933-e84012a29211"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.991983 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-config" (OuterVolumeSpecName: "config") pod "fa92cf8a-56c1-4b38-889c-23a578b74b90" (UID: "fa92cf8a-56c1-4b38-889c-23a578b74b90"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.992606 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-client-ca" (OuterVolumeSpecName: "client-ca") pod "fa92cf8a-56c1-4b38-889c-23a578b74b90" (UID: "fa92cf8a-56c1-4b38-889c-23a578b74b90"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.992967 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-client-ca" (OuterVolumeSpecName: "client-ca") pod "b9b65044-9326-4a72-a933-e84012a29211" (UID: "b9b65044-9326-4a72-a933-e84012a29211"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:05:09 crc kubenswrapper[4956]: I1126 17:05:09.993924 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-config" (OuterVolumeSpecName: "config") pod "b9b65044-9326-4a72-a933-e84012a29211" (UID: "b9b65044-9326-4a72-a933-e84012a29211"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.000421 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9b65044-9326-4a72-a933-e84012a29211-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "b9b65044-9326-4a72-a933-e84012a29211" (UID: "b9b65044-9326-4a72-a933-e84012a29211"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.005078 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b65044-9326-4a72-a933-e84012a29211-kube-api-access-vxsrf" (OuterVolumeSpecName: "kube-api-access-vxsrf") pod "b9b65044-9326-4a72-a933-e84012a29211" (UID: "b9b65044-9326-4a72-a933-e84012a29211"). InnerVolumeSpecName "kube-api-access-vxsrf". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.016115 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fa92cf8a-56c1-4b38-889c-23a578b74b90-kube-api-access-q57kt" (OuterVolumeSpecName: "kube-api-access-q57kt") pod "fa92cf8a-56c1-4b38-889c-23a578b74b90" (UID: "fa92cf8a-56c1-4b38-889c-23a578b74b90"). InnerVolumeSpecName "kube-api-access-q57kt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.022000 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fa92cf8a-56c1-4b38-889c-23a578b74b90-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "fa92cf8a-56c1-4b38-889c-23a578b74b90" (UID: "fa92cf8a-56c1-4b38-889c-23a578b74b90"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.061890 4956 scope.go:117] "RemoveContainer" containerID="932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda" Nov 26 17:05:10 crc kubenswrapper[4956]: E1126 17:05:10.062751 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda\": container with ID starting with 932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda not found: ID does not exist" containerID="932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.062849 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda"} err="failed to get container status \"932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda\": rpc error: code = NotFound desc = could not find container \"932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda\": container with ID starting with 932eb40dc8a29ec1560570ad0680094c6d132e4254e74a98bcbbaea31fffbdda not found: ID does not exist" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090808 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/fa92cf8a-56c1-4b38-889c-23a578b74b90-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090848 4956 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090858 4956 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-client-ca\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090885 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q57kt\" (UniqueName: \"kubernetes.io/projected/fa92cf8a-56c1-4b38-889c-23a578b74b90-kube-api-access-q57kt\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090900 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vxsrf\" (UniqueName: \"kubernetes.io/projected/b9b65044-9326-4a72-a933-e84012a29211-kube-api-access-vxsrf\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090911 4956 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090921 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b65044-9326-4a72-a933-e84012a29211-config\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090930 4956 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/fa92cf8a-56c1-4b38-889c-23a578b74b90-config\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.090938 4956 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b9b65044-9326-4a72-a933-e84012a29211-serving-cert\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.300081 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v5xlx"] Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.303476 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-v5xlx"] Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.315965 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2"] Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.321857 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-kr7n2"] Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373254 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2"] Nov 26 17:05:10 crc kubenswrapper[4956]: E1126 17:05:10.373582 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="21d7cda2-cd2b-4ad8-911f-e117e129381f" containerName="registry-server" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373600 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="21d7cda2-cd2b-4ad8-911f-e117e129381f" containerName="registry-server" Nov 26 17:05:10 crc kubenswrapper[4956]: E1126 17:05:10.373626 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b65044-9326-4a72-a933-e84012a29211" containerName="controller-manager" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373633 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b65044-9326-4a72-a933-e84012a29211" containerName="controller-manager" Nov 26 17:05:10 crc kubenswrapper[4956]: E1126 17:05:10.373645 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerName="route-controller-manager" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373653 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerName="route-controller-manager" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373769 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" containerName="route-controller-manager" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373785 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b65044-9326-4a72-a933-e84012a29211" containerName="controller-manager" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.373801 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="21d7cda2-cd2b-4ad8-911f-e117e129381f" containerName="registry-server" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.374703 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.376474 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j97nd" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.380237 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2"] Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.498903 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.499007 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.499043 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2vrj6\" (UniqueName: \"kubernetes.io/projected/172f2407-1798-487a-b864-27f28d867ecd-kube-api-access-2vrj6\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.601192 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.601252 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.601290 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2vrj6\" (UniqueName: \"kubernetes.io/projected/172f2407-1798-487a-b864-27f28d867ecd-kube-api-access-2vrj6\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.601853 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-bundle\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.602109 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-util\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.619137 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2vrj6\" (UniqueName: \"kubernetes.io/projected/172f2407-1798-487a-b864-27f28d867ecd-kube-api-access-2vrj6\") pod \"27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.703384 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.929066 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2"] Nov 26 17:05:10 crc kubenswrapper[4956]: W1126 17:05:10.940503 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod172f2407_1798_487a_b864_27f28d867ecd.slice/crio-0cf0bd48d1b85b7f553677ba556210025dad3c90f1f21b72a6986525eff51aaf WatchSource:0}: Error finding container 0cf0bd48d1b85b7f553677ba556210025dad3c90f1f21b72a6986525eff51aaf: Status 404 returned error can't find the container with id 0cf0bd48d1b85b7f553677ba556210025dad3c90f1f21b72a6986525eff51aaf Nov 26 17:05:10 crc kubenswrapper[4956]: I1126 17:05:10.979022 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" event={"ID":"172f2407-1798-487a-b864-27f28d867ecd","Type":"ContainerStarted","Data":"0cf0bd48d1b85b7f553677ba556210025dad3c90f1f21b72a6986525eff51aaf"} Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.001790 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b65044-9326-4a72-a933-e84012a29211" path="/var/lib/kubelet/pods/b9b65044-9326-4a72-a933-e84012a29211/volumes" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.002941 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fa92cf8a-56c1-4b38-889c-23a578b74b90" path="/var/lib/kubelet/pods/fa92cf8a-56c1-4b38-889c-23a578b74b90/volumes" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.511683 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds"] Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.513060 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.516978 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.517114 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.517219 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.517256 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.518530 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.518532 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.518555 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx"] Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.519681 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.523306 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.523550 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.523709 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.523896 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.524003 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.528970 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx"] Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.536451 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.536731 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.554167 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds"] Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615762 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-config\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615807 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-proxy-ca-bundles\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615849 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afbcf363-8916-4d31-98d7-fcae22eda7dc-config\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615885 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/afbcf363-8916-4d31-98d7-fcae22eda7dc-client-ca\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615915 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpjk5\" (UniqueName: \"kubernetes.io/projected/730fc797-883f-4f01-b32c-f1190a4673bc-kube-api-access-xpjk5\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615947 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afbcf363-8916-4d31-98d7-fcae22eda7dc-serving-cert\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615973 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/730fc797-883f-4f01-b32c-f1190a4673bc-serving-cert\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.615992 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-client-ca\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.616025 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmjbk\" (UniqueName: \"kubernetes.io/projected/afbcf363-8916-4d31-98d7-fcae22eda7dc-kube-api-access-hmjbk\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717077 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmjbk\" (UniqueName: \"kubernetes.io/projected/afbcf363-8916-4d31-98d7-fcae22eda7dc-kube-api-access-hmjbk\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717166 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-config\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717190 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-proxy-ca-bundles\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717225 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afbcf363-8916-4d31-98d7-fcae22eda7dc-config\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717247 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/afbcf363-8916-4d31-98d7-fcae22eda7dc-client-ca\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717270 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpjk5\" (UniqueName: \"kubernetes.io/projected/730fc797-883f-4f01-b32c-f1190a4673bc-kube-api-access-xpjk5\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717304 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afbcf363-8916-4d31-98d7-fcae22eda7dc-serving-cert\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717331 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/730fc797-883f-4f01-b32c-f1190a4673bc-serving-cert\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.717352 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-client-ca\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.718757 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/afbcf363-8916-4d31-98d7-fcae22eda7dc-client-ca\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.718920 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-client-ca\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.719049 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-proxy-ca-bundles\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.719279 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/730fc797-883f-4f01-b32c-f1190a4673bc-config\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.719783 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/afbcf363-8916-4d31-98d7-fcae22eda7dc-config\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.728107 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/afbcf363-8916-4d31-98d7-fcae22eda7dc-serving-cert\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.729167 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/730fc797-883f-4f01-b32c-f1190a4673bc-serving-cert\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.738752 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmjbk\" (UniqueName: \"kubernetes.io/projected/afbcf363-8916-4d31-98d7-fcae22eda7dc-kube-api-access-hmjbk\") pod \"route-controller-manager-7cbbbb987f-hnmtx\" (UID: \"afbcf363-8916-4d31-98d7-fcae22eda7dc\") " pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.740339 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpjk5\" (UniqueName: \"kubernetes.io/projected/730fc797-883f-4f01-b32c-f1190a4673bc-kube-api-access-xpjk5\") pod \"controller-manager-5f99bb8fb9-4vpds\" (UID: \"730fc797-883f-4f01-b32c-f1190a4673bc\") " pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.844662 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.856504 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:11 crc kubenswrapper[4956]: I1126 17:05:11.998692 4956 generic.go:334] "Generic (PLEG): container finished" podID="172f2407-1798-487a-b864-27f28d867ecd" containerID="e9a030d42e0cbcc920d00e421ac2a21aa3c164d61c8b9f829b790d6f58ab52d0" exitCode=0 Nov 26 17:05:12 crc kubenswrapper[4956]: I1126 17:05:11.999905 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" event={"ID":"172f2407-1798-487a-b864-27f28d867ecd","Type":"ContainerDied","Data":"e9a030d42e0cbcc920d00e421ac2a21aa3c164d61c8b9f829b790d6f58ab52d0"} Nov 26 17:05:12 crc kubenswrapper[4956]: I1126 17:05:12.144081 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx"] Nov 26 17:05:12 crc kubenswrapper[4956]: W1126 17:05:12.151516 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podafbcf363_8916_4d31_98d7_fcae22eda7dc.slice/crio-ace63a4a05878cf611d5a65c8e114f092874e34e0a1791bc16d8cd138e7e7229 WatchSource:0}: Error finding container ace63a4a05878cf611d5a65c8e114f092874e34e0a1791bc16d8cd138e7e7229: Status 404 returned error can't find the container with id ace63a4a05878cf611d5a65c8e114f092874e34e0a1791bc16d8cd138e7e7229 Nov 26 17:05:12 crc kubenswrapper[4956]: I1126 17:05:12.193258 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds"] Nov 26 17:05:12 crc kubenswrapper[4956]: W1126 17:05:12.200612 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod730fc797_883f_4f01_b32c_f1190a4673bc.slice/crio-793d2e6ae76fc05559450bcacbd91ec39962de36d2a858d066074f291770008c WatchSource:0}: Error finding container 793d2e6ae76fc05559450bcacbd91ec39962de36d2a858d066074f291770008c: Status 404 returned error can't find the container with id 793d2e6ae76fc05559450bcacbd91ec39962de36d2a858d066074f291770008c Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.011829 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.012360 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" event={"ID":"730fc797-883f-4f01-b32c-f1190a4673bc","Type":"ContainerStarted","Data":"13e8130498aeaa192a510cb94f6c96dbb60f9efa159c8987aa829bd91e9af069"} Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.012381 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" event={"ID":"730fc797-883f-4f01-b32c-f1190a4673bc","Type":"ContainerStarted","Data":"793d2e6ae76fc05559450bcacbd91ec39962de36d2a858d066074f291770008c"} Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.014016 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" event={"ID":"afbcf363-8916-4d31-98d7-fcae22eda7dc","Type":"ContainerStarted","Data":"3836f99cd39cb3fb7ba1520f2286ead6df45c9e2032164949e076def50ce9064"} Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.014063 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" event={"ID":"afbcf363-8916-4d31-98d7-fcae22eda7dc","Type":"ContainerStarted","Data":"ace63a4a05878cf611d5a65c8e114f092874e34e0a1791bc16d8cd138e7e7229"} Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.014522 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.023179 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.082153 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-5f99bb8fb9-4vpds" podStartSLOduration=4.082128405 podStartE2EDuration="4.082128405s" podCreationTimestamp="2025-11-26 17:05:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:05:13.045124998 +0000 UTC m=+758.741085570" watchObservedRunningTime="2025-11-26 17:05:13.082128405 +0000 UTC m=+758.778088957" Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.287572 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" Nov 26 17:05:13 crc kubenswrapper[4956]: I1126 17:05:13.314927 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-7cbbbb987f-hnmtx" podStartSLOduration=4.314897845 podStartE2EDuration="4.314897845s" podCreationTimestamp="2025-11-26 17:05:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:05:13.129075876 +0000 UTC m=+758.825036438" watchObservedRunningTime="2025-11-26 17:05:13.314897845 +0000 UTC m=+759.010858407" Nov 26 17:05:14 crc kubenswrapper[4956]: I1126 17:05:14.021275 4956 generic.go:334] "Generic (PLEG): container finished" podID="172f2407-1798-487a-b864-27f28d867ecd" containerID="d84dd13645f695c7c50aff2462f97c66db2853b727ccb1e4350f830dd7fa326e" exitCode=0 Nov 26 17:05:14 crc kubenswrapper[4956]: I1126 17:05:14.022497 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" event={"ID":"172f2407-1798-487a-b864-27f28d867ecd","Type":"ContainerDied","Data":"d84dd13645f695c7c50aff2462f97c66db2853b727ccb1e4350f830dd7fa326e"} Nov 26 17:05:15 crc kubenswrapper[4956]: I1126 17:05:15.031775 4956 generic.go:334] "Generic (PLEG): container finished" podID="172f2407-1798-487a-b864-27f28d867ecd" containerID="01158968ded21d0b6a415277d23d8f8fae07dca88d84276c8575d0d1f31c3a9b" exitCode=0 Nov 26 17:05:15 crc kubenswrapper[4956]: I1126 17:05:15.031961 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" event={"ID":"172f2407-1798-487a-b864-27f28d867ecd","Type":"ContainerDied","Data":"01158968ded21d0b6a415277d23d8f8fae07dca88d84276c8575d0d1f31c3a9b"} Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.393577 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.496643 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-util\") pod \"172f2407-1798-487a-b864-27f28d867ecd\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.496827 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-bundle\") pod \"172f2407-1798-487a-b864-27f28d867ecd\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.496966 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2vrj6\" (UniqueName: \"kubernetes.io/projected/172f2407-1798-487a-b864-27f28d867ecd-kube-api-access-2vrj6\") pod \"172f2407-1798-487a-b864-27f28d867ecd\" (UID: \"172f2407-1798-487a-b864-27f28d867ecd\") " Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.498518 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-bundle" (OuterVolumeSpecName: "bundle") pod "172f2407-1798-487a-b864-27f28d867ecd" (UID: "172f2407-1798-487a-b864-27f28d867ecd"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.504526 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/172f2407-1798-487a-b864-27f28d867ecd-kube-api-access-2vrj6" (OuterVolumeSpecName: "kube-api-access-2vrj6") pod "172f2407-1798-487a-b864-27f28d867ecd" (UID: "172f2407-1798-487a-b864-27f28d867ecd"). InnerVolumeSpecName "kube-api-access-2vrj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.511083 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-util" (OuterVolumeSpecName: "util") pod "172f2407-1798-487a-b864-27f28d867ecd" (UID: "172f2407-1798-487a-b864-27f28d867ecd"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.598947 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.599363 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/172f2407-1798-487a-b864-27f28d867ecd-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:16 crc kubenswrapper[4956]: I1126 17:05:16.599479 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2vrj6\" (UniqueName: \"kubernetes.io/projected/172f2407-1798-487a-b864-27f28d867ecd-kube-api-access-2vrj6\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:17 crc kubenswrapper[4956]: I1126 17:05:17.051035 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" event={"ID":"172f2407-1798-487a-b864-27f28d867ecd","Type":"ContainerDied","Data":"0cf0bd48d1b85b7f553677ba556210025dad3c90f1f21b72a6986525eff51aaf"} Nov 26 17:05:17 crc kubenswrapper[4956]: I1126 17:05:17.051100 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0cf0bd48d1b85b7f553677ba556210025dad3c90f1f21b72a6986525eff51aaf" Nov 26 17:05:17 crc kubenswrapper[4956]: I1126 17:05:17.051612 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2" Nov 26 17:05:17 crc kubenswrapper[4956]: I1126 17:05:17.477244 4956 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.134493 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c"] Nov 26 17:05:24 crc kubenswrapper[4956]: E1126 17:05:24.135635 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="pull" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.135654 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="pull" Nov 26 17:05:24 crc kubenswrapper[4956]: E1126 17:05:24.135675 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="extract" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.135682 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="extract" Nov 26 17:05:24 crc kubenswrapper[4956]: E1126 17:05:24.135697 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="util" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.135704 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="util" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.135813 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="172f2407-1798-487a-b864-27f28d867ecd" containerName="extract" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.136296 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.138506 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.138593 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.143972 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tl4dj" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.153834 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c"] Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.228229 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-webhook-cert\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.228293 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w59p9\" (UniqueName: \"kubernetes.io/projected/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-kube-api-access-w59p9\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.228360 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-apiservice-cert\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.329355 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-apiservice-cert\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.329423 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-webhook-cert\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.329468 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w59p9\" (UniqueName: \"kubernetes.io/projected/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-kube-api-access-w59p9\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.339751 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-webhook-cert\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.339841 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-apiservice-cert\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.347606 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w59p9\" (UniqueName: \"kubernetes.io/projected/91276233-6cc3-408e-b9e4-e8a4c7ecc60d-kube-api-access-w59p9\") pod \"mariadb-operator-controller-manager-7674ccc9cf-j5f9c\" (UID: \"91276233-6cc3-408e-b9e4-e8a4c7ecc60d\") " pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.496415 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:24 crc kubenswrapper[4956]: I1126 17:05:24.817719 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c"] Nov 26 17:05:24 crc kubenswrapper[4956]: W1126 17:05:24.828455 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod91276233_6cc3_408e_b9e4_e8a4c7ecc60d.slice/crio-8ea359c2938bac7e69afd0c5b96d1a2c779ee4a4379f1c2424d41a3185fca3eb WatchSource:0}: Error finding container 8ea359c2938bac7e69afd0c5b96d1a2c779ee4a4379f1c2424d41a3185fca3eb: Status 404 returned error can't find the container with id 8ea359c2938bac7e69afd0c5b96d1a2c779ee4a4379f1c2424d41a3185fca3eb Nov 26 17:05:25 crc kubenswrapper[4956]: I1126 17:05:25.101840 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" event={"ID":"91276233-6cc3-408e-b9e4-e8a4c7ecc60d","Type":"ContainerStarted","Data":"8ea359c2938bac7e69afd0c5b96d1a2c779ee4a4379f1c2424d41a3185fca3eb"} Nov 26 17:05:29 crc kubenswrapper[4956]: I1126 17:05:29.135914 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" event={"ID":"91276233-6cc3-408e-b9e4-e8a4c7ecc60d","Type":"ContainerStarted","Data":"fa4d811b64502f501dbebbb3bd74b87ace4ad7364aa8b664e8656f59082be9b1"} Nov 26 17:05:29 crc kubenswrapper[4956]: I1126 17:05:29.137251 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:29 crc kubenswrapper[4956]: I1126 17:05:29.167304 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" podStartSLOduration=1.52271844 podStartE2EDuration="5.167277017s" podCreationTimestamp="2025-11-26 17:05:24 +0000 UTC" firstStartedPulling="2025-11-26 17:05:24.830164528 +0000 UTC m=+770.526125080" lastFinishedPulling="2025-11-26 17:05:28.474723105 +0000 UTC m=+774.170683657" observedRunningTime="2025-11-26 17:05:29.166048492 +0000 UTC m=+774.862009074" watchObservedRunningTime="2025-11-26 17:05:29.167277017 +0000 UTC m=+774.863237599" Nov 26 17:05:34 crc kubenswrapper[4956]: I1126 17:05:34.502621 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.459185 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qfg67"] Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.462147 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.468309 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfg67"] Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.543239 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-catalog-content\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.543348 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m9fjr\" (UniqueName: \"kubernetes.io/projected/df02f319-b6aa-4ed1-87d7-36da32e41700-kube-api-access-m9fjr\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.543436 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-utilities\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.644489 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-catalog-content\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.644560 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m9fjr\" (UniqueName: \"kubernetes.io/projected/df02f319-b6aa-4ed1-87d7-36da32e41700-kube-api-access-m9fjr\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.644598 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-utilities\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.645205 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-utilities\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.645207 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-catalog-content\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.675427 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m9fjr\" (UniqueName: \"kubernetes.io/projected/df02f319-b6aa-4ed1-87d7-36da32e41700-kube-api-access-m9fjr\") pod \"redhat-marketplace-qfg67\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:37 crc kubenswrapper[4956]: I1126 17:05:37.840838 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:38 crc kubenswrapper[4956]: I1126 17:05:38.341395 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfg67"] Nov 26 17:05:38 crc kubenswrapper[4956]: W1126 17:05:38.353641 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddf02f319_b6aa_4ed1_87d7_36da32e41700.slice/crio-1029a52a4482b520963379f59a8ccbc018810cebe5b964b648921a17420be7b9 WatchSource:0}: Error finding container 1029a52a4482b520963379f59a8ccbc018810cebe5b964b648921a17420be7b9: Status 404 returned error can't find the container with id 1029a52a4482b520963379f59a8ccbc018810cebe5b964b648921a17420be7b9 Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.194892 4956 generic.go:334] "Generic (PLEG): container finished" podID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerID="1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6" exitCode=0 Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.194974 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfg67" event={"ID":"df02f319-b6aa-4ed1-87d7-36da32e41700","Type":"ContainerDied","Data":"1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6"} Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.195306 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfg67" event={"ID":"df02f319-b6aa-4ed1-87d7-36da32e41700","Type":"ContainerStarted","Data":"1029a52a4482b520963379f59a8ccbc018810cebe5b964b648921a17420be7b9"} Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.552290 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.552392 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.552504 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.553652 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"60fb36ae8e9dbdf5a9ddd39194e72f5c72fcc2005c642dd846c7d001b368d385"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.553807 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://60fb36ae8e9dbdf5a9ddd39194e72f5c72fcc2005c642dd846c7d001b368d385" gracePeriod=600 Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.848113 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-index-t6ctx"] Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.849307 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.851213 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-2htkq" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.858133 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-t6ctx"] Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.874012 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sz6zs\" (UniqueName: \"kubernetes.io/projected/24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af-kube-api-access-sz6zs\") pod \"infra-operator-index-t6ctx\" (UID: \"24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af\") " pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.975524 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sz6zs\" (UniqueName: \"kubernetes.io/projected/24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af-kube-api-access-sz6zs\") pod \"infra-operator-index-t6ctx\" (UID: \"24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af\") " pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:39 crc kubenswrapper[4956]: I1126 17:05:39.996334 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sz6zs\" (UniqueName: \"kubernetes.io/projected/24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af-kube-api-access-sz6zs\") pod \"infra-operator-index-t6ctx\" (UID: \"24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af\") " pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:40 crc kubenswrapper[4956]: I1126 17:05:40.176286 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:40 crc kubenswrapper[4956]: I1126 17:05:40.206503 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="60fb36ae8e9dbdf5a9ddd39194e72f5c72fcc2005c642dd846c7d001b368d385" exitCode=0 Nov 26 17:05:40 crc kubenswrapper[4956]: I1126 17:05:40.206558 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"60fb36ae8e9dbdf5a9ddd39194e72f5c72fcc2005c642dd846c7d001b368d385"} Nov 26 17:05:40 crc kubenswrapper[4956]: I1126 17:05:40.206606 4956 scope.go:117] "RemoveContainer" containerID="abea58a965d35521fe5183bcbe17e6d6a5bcb02b082905c716d7ab2a2436c54e" Nov 26 17:05:40 crc kubenswrapper[4956]: I1126 17:05:40.642374 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-index-t6ctx"] Nov 26 17:05:40 crc kubenswrapper[4956]: W1126 17:05:40.720157 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod24c5dfcb_c59d_4acf_b0c2_3d0f9d6f12af.slice/crio-1cf6d9361d2eb8b2d67724203462a17a9d9e55fcd68b7119831f1b861d9e4eae WatchSource:0}: Error finding container 1cf6d9361d2eb8b2d67724203462a17a9d9e55fcd68b7119831f1b861d9e4eae: Status 404 returned error can't find the container with id 1cf6d9361d2eb8b2d67724203462a17a9d9e55fcd68b7119831f1b861d9e4eae Nov 26 17:05:41 crc kubenswrapper[4956]: I1126 17:05:41.213133 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-t6ctx" event={"ID":"24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af","Type":"ContainerStarted","Data":"1cf6d9361d2eb8b2d67724203462a17a9d9e55fcd68b7119831f1b861d9e4eae"} Nov 26 17:05:41 crc kubenswrapper[4956]: I1126 17:05:41.214728 4956 generic.go:334] "Generic (PLEG): container finished" podID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerID="9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289" exitCode=0 Nov 26 17:05:41 crc kubenswrapper[4956]: I1126 17:05:41.214798 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfg67" event={"ID":"df02f319-b6aa-4ed1-87d7-36da32e41700","Type":"ContainerDied","Data":"9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289"} Nov 26 17:05:41 crc kubenswrapper[4956]: I1126 17:05:41.216797 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"0683eafa4a97fe7cebdb5b35d5b4a8e27469c373b19946cd7cb9d38e21f774da"} Nov 26 17:05:42 crc kubenswrapper[4956]: I1126 17:05:42.232808 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-index-t6ctx" event={"ID":"24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af","Type":"ContainerStarted","Data":"8a0c5889218941be68b327b8bac782550ed2cd6b2fb313405b59b008aed49708"} Nov 26 17:05:42 crc kubenswrapper[4956]: I1126 17:05:42.253784 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-index-t6ctx" podStartSLOduration=2.107264667 podStartE2EDuration="3.253765492s" podCreationTimestamp="2025-11-26 17:05:39 +0000 UTC" firstStartedPulling="2025-11-26 17:05:40.723347612 +0000 UTC m=+786.419308164" lastFinishedPulling="2025-11-26 17:05:41.869848397 +0000 UTC m=+787.565808989" observedRunningTime="2025-11-26 17:05:42.252041543 +0000 UTC m=+787.948002105" watchObservedRunningTime="2025-11-26 17:05:42.253765492 +0000 UTC m=+787.949726044" Nov 26 17:05:43 crc kubenswrapper[4956]: I1126 17:05:43.242331 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfg67" event={"ID":"df02f319-b6aa-4ed1-87d7-36da32e41700","Type":"ContainerStarted","Data":"a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67"} Nov 26 17:05:43 crc kubenswrapper[4956]: I1126 17:05:43.265892 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qfg67" podStartSLOduration=2.736978461 podStartE2EDuration="6.265855145s" podCreationTimestamp="2025-11-26 17:05:37 +0000 UTC" firstStartedPulling="2025-11-26 17:05:39.197032901 +0000 UTC m=+784.892993453" lastFinishedPulling="2025-11-26 17:05:42.725909585 +0000 UTC m=+788.421870137" observedRunningTime="2025-11-26 17:05:43.261825179 +0000 UTC m=+788.957785811" watchObservedRunningTime="2025-11-26 17:05:43.265855145 +0000 UTC m=+788.961815697" Nov 26 17:05:47 crc kubenswrapper[4956]: I1126 17:05:47.842038 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:47 crc kubenswrapper[4956]: I1126 17:05:47.842735 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:47 crc kubenswrapper[4956]: I1126 17:05:47.895670 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:48 crc kubenswrapper[4956]: I1126 17:05:48.321933 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:49 crc kubenswrapper[4956]: I1126 17:05:49.642572 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfg67"] Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.177323 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.177399 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.215009 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.296590 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qfg67" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="registry-server" containerID="cri-o://a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67" gracePeriod=2 Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.335935 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-index-t6ctx" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.750730 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.830988 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-utilities\") pod \"df02f319-b6aa-4ed1-87d7-36da32e41700\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.831313 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-catalog-content\") pod \"df02f319-b6aa-4ed1-87d7-36da32e41700\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.831352 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m9fjr\" (UniqueName: \"kubernetes.io/projected/df02f319-b6aa-4ed1-87d7-36da32e41700-kube-api-access-m9fjr\") pod \"df02f319-b6aa-4ed1-87d7-36da32e41700\" (UID: \"df02f319-b6aa-4ed1-87d7-36da32e41700\") " Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.837221 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-utilities" (OuterVolumeSpecName: "utilities") pod "df02f319-b6aa-4ed1-87d7-36da32e41700" (UID: "df02f319-b6aa-4ed1-87d7-36da32e41700"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.856141 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df02f319-b6aa-4ed1-87d7-36da32e41700-kube-api-access-m9fjr" (OuterVolumeSpecName: "kube-api-access-m9fjr") pod "df02f319-b6aa-4ed1-87d7-36da32e41700" (UID: "df02f319-b6aa-4ed1-87d7-36da32e41700"). InnerVolumeSpecName "kube-api-access-m9fjr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.877701 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "df02f319-b6aa-4ed1-87d7-36da32e41700" (UID: "df02f319-b6aa-4ed1-87d7-36da32e41700"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.932941 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.933005 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m9fjr\" (UniqueName: \"kubernetes.io/projected/df02f319-b6aa-4ed1-87d7-36da32e41700-kube-api-access-m9fjr\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:50 crc kubenswrapper[4956]: I1126 17:05:50.933030 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/df02f319-b6aa-4ed1-87d7-36da32e41700-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.305841 4956 generic.go:334] "Generic (PLEG): container finished" podID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerID="a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67" exitCode=0 Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.305920 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfg67" event={"ID":"df02f319-b6aa-4ed1-87d7-36da32e41700","Type":"ContainerDied","Data":"a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67"} Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.305988 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qfg67" event={"ID":"df02f319-b6aa-4ed1-87d7-36da32e41700","Type":"ContainerDied","Data":"1029a52a4482b520963379f59a8ccbc018810cebe5b964b648921a17420be7b9"} Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.305995 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qfg67" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.306012 4956 scope.go:117] "RemoveContainer" containerID="a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.331400 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfg67"] Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.332241 4956 scope.go:117] "RemoveContainer" containerID="9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.333224 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qfg67"] Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.349256 4956 scope.go:117] "RemoveContainer" containerID="1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.369148 4956 scope.go:117] "RemoveContainer" containerID="a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67" Nov 26 17:05:51 crc kubenswrapper[4956]: E1126 17:05:51.370173 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67\": container with ID starting with a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67 not found: ID does not exist" containerID="a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.370230 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67"} err="failed to get container status \"a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67\": rpc error: code = NotFound desc = could not find container \"a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67\": container with ID starting with a2ed82d718d06948a005aaa902b026e2e10de69e80aa53608beeb85a3dfe3a67 not found: ID does not exist" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.370274 4956 scope.go:117] "RemoveContainer" containerID="9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289" Nov 26 17:05:51 crc kubenswrapper[4956]: E1126 17:05:51.370622 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289\": container with ID starting with 9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289 not found: ID does not exist" containerID="9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.370662 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289"} err="failed to get container status \"9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289\": rpc error: code = NotFound desc = could not find container \"9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289\": container with ID starting with 9edb210243b5bfae49e66b7b0786f482ae1eb9119c7f46e420d6b44d1d1a1289 not found: ID does not exist" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.370684 4956 scope.go:117] "RemoveContainer" containerID="1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6" Nov 26 17:05:51 crc kubenswrapper[4956]: E1126 17:05:51.371040 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6\": container with ID starting with 1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6 not found: ID does not exist" containerID="1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6" Nov 26 17:05:51 crc kubenswrapper[4956]: I1126 17:05:51.371078 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6"} err="failed to get container status \"1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6\": rpc error: code = NotFound desc = could not find container \"1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6\": container with ID starting with 1d3b932dbb26f8e96e3ee0901f5ddde7a81a59b05e4474c96d4c8af08cfb59e6 not found: ID does not exist" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.513030 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq"] Nov 26 17:05:52 crc kubenswrapper[4956]: E1126 17:05:52.513495 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="extract-utilities" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.513526 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="extract-utilities" Nov 26 17:05:52 crc kubenswrapper[4956]: E1126 17:05:52.513560 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="registry-server" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.513572 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="registry-server" Nov 26 17:05:52 crc kubenswrapper[4956]: E1126 17:05:52.513593 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="extract-content" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.513604 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="extract-content" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.513887 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" containerName="registry-server" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.515572 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.517750 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq"] Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.557188 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j97nd" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.658168 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/495cf683-8b21-4b50-9207-b7727070c747-kube-api-access-dvlr7\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.659217 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.659503 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.761312 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/495cf683-8b21-4b50-9207-b7727070c747-kube-api-access-dvlr7\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.761435 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.761478 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.762105 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-bundle\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.762574 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-util\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.785056 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/495cf683-8b21-4b50-9207-b7727070c747-kube-api-access-dvlr7\") pod \"5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:52 crc kubenswrapper[4956]: I1126 17:05:52.877465 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:53 crc kubenswrapper[4956]: I1126 17:05:53.010905 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df02f319-b6aa-4ed1-87d7-36da32e41700" path="/var/lib/kubelet/pods/df02f319-b6aa-4ed1-87d7-36da32e41700/volumes" Nov 26 17:05:53 crc kubenswrapper[4956]: I1126 17:05:53.314259 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq"] Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.052803 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rwggh"] Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.055112 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.073556 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rwggh"] Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.081014 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-utilities\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.081087 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q98ms\" (UniqueName: \"kubernetes.io/projected/0b59c522-cd4b-4615-b579-9a5574cc72d4-kube-api-access-q98ms\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.081125 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-catalog-content\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.182161 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-utilities\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.182235 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q98ms\" (UniqueName: \"kubernetes.io/projected/0b59c522-cd4b-4615-b579-9a5574cc72d4-kube-api-access-q98ms\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.182271 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-catalog-content\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.182793 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-catalog-content\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.183057 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-utilities\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.228723 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q98ms\" (UniqueName: \"kubernetes.io/projected/0b59c522-cd4b-4615-b579-9a5574cc72d4-kube-api-access-q98ms\") pod \"certified-operators-rwggh\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.340965 4956 generic.go:334] "Generic (PLEG): container finished" podID="495cf683-8b21-4b50-9207-b7727070c747" containerID="15e3ea659d6154ef8a6200b58e42243efaf9ae4af16af6a49051febaef362ea0" exitCode=0 Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.341018 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" event={"ID":"495cf683-8b21-4b50-9207-b7727070c747","Type":"ContainerDied","Data":"15e3ea659d6154ef8a6200b58e42243efaf9ae4af16af6a49051febaef362ea0"} Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.341045 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" event={"ID":"495cf683-8b21-4b50-9207-b7727070c747","Type":"ContainerStarted","Data":"adc45d9de082b37936a929ad07ba121f8e63ccb7fb53d51fc78dd3f80548f002"} Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.371352 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:05:54 crc kubenswrapper[4956]: I1126 17:05:54.676605 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rwggh"] Nov 26 17:05:54 crc kubenswrapper[4956]: W1126 17:05:54.688989 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0b59c522_cd4b_4615_b579_9a5574cc72d4.slice/crio-b29238e51d7e6a8787dca95a05d61b88a4978be5148e88956320095365e757b8 WatchSource:0}: Error finding container b29238e51d7e6a8787dca95a05d61b88a4978be5148e88956320095365e757b8: Status 404 returned error can't find the container with id b29238e51d7e6a8787dca95a05d61b88a4978be5148e88956320095365e757b8 Nov 26 17:05:55 crc kubenswrapper[4956]: I1126 17:05:55.350194 4956 generic.go:334] "Generic (PLEG): container finished" podID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerID="95ed27c11c47f43a0c7e6887d93ed38f50d639fffc35a53b0c7be109c23bec0f" exitCode=0 Nov 26 17:05:55 crc kubenswrapper[4956]: I1126 17:05:55.350256 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerDied","Data":"95ed27c11c47f43a0c7e6887d93ed38f50d639fffc35a53b0c7be109c23bec0f"} Nov 26 17:05:55 crc kubenswrapper[4956]: I1126 17:05:55.350672 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerStarted","Data":"b29238e51d7e6a8787dca95a05d61b88a4978be5148e88956320095365e757b8"} Nov 26 17:05:56 crc kubenswrapper[4956]: I1126 17:05:56.358528 4956 generic.go:334] "Generic (PLEG): container finished" podID="495cf683-8b21-4b50-9207-b7727070c747" containerID="e668b53330ccf7b127995341c655a6911e3b60da5e39500039f9ed2350ac5067" exitCode=0 Nov 26 17:05:56 crc kubenswrapper[4956]: I1126 17:05:56.358597 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" event={"ID":"495cf683-8b21-4b50-9207-b7727070c747","Type":"ContainerDied","Data":"e668b53330ccf7b127995341c655a6911e3b60da5e39500039f9ed2350ac5067"} Nov 26 17:05:56 crc kubenswrapper[4956]: I1126 17:05:56.362285 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerStarted","Data":"d1e16f6d33694dd4cb6a8be8ec12fbb14f6ca44980d487fd4c04338bafcbb701"} Nov 26 17:05:57 crc kubenswrapper[4956]: I1126 17:05:57.370311 4956 generic.go:334] "Generic (PLEG): container finished" podID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerID="d1e16f6d33694dd4cb6a8be8ec12fbb14f6ca44980d487fd4c04338bafcbb701" exitCode=0 Nov 26 17:05:57 crc kubenswrapper[4956]: I1126 17:05:57.370388 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerDied","Data":"d1e16f6d33694dd4cb6a8be8ec12fbb14f6ca44980d487fd4c04338bafcbb701"} Nov 26 17:05:57 crc kubenswrapper[4956]: I1126 17:05:57.373663 4956 generic.go:334] "Generic (PLEG): container finished" podID="495cf683-8b21-4b50-9207-b7727070c747" containerID="85ac8ad266af8bcd15bde458332650f71baecce711279351c84944a4e32fb991" exitCode=0 Nov 26 17:05:57 crc kubenswrapper[4956]: I1126 17:05:57.373734 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" event={"ID":"495cf683-8b21-4b50-9207-b7727070c747","Type":"ContainerDied","Data":"85ac8ad266af8bcd15bde458332650f71baecce711279351c84944a4e32fb991"} Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.684722 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.852103 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-bundle\") pod \"495cf683-8b21-4b50-9207-b7727070c747\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.852161 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/495cf683-8b21-4b50-9207-b7727070c747-kube-api-access-dvlr7\") pod \"495cf683-8b21-4b50-9207-b7727070c747\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.852195 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-util\") pod \"495cf683-8b21-4b50-9207-b7727070c747\" (UID: \"495cf683-8b21-4b50-9207-b7727070c747\") " Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.853564 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-bundle" (OuterVolumeSpecName: "bundle") pod "495cf683-8b21-4b50-9207-b7727070c747" (UID: "495cf683-8b21-4b50-9207-b7727070c747"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.859087 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/495cf683-8b21-4b50-9207-b7727070c747-kube-api-access-dvlr7" (OuterVolumeSpecName: "kube-api-access-dvlr7") pod "495cf683-8b21-4b50-9207-b7727070c747" (UID: "495cf683-8b21-4b50-9207-b7727070c747"). InnerVolumeSpecName "kube-api-access-dvlr7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.943646 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-util" (OuterVolumeSpecName: "util") pod "495cf683-8b21-4b50-9207-b7727070c747" (UID: "495cf683-8b21-4b50-9207-b7727070c747"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.953645 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.953690 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvlr7\" (UniqueName: \"kubernetes.io/projected/495cf683-8b21-4b50-9207-b7727070c747-kube-api-access-dvlr7\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:58 crc kubenswrapper[4956]: I1126 17:05:58.953712 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/495cf683-8b21-4b50-9207-b7727070c747-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:05:59 crc kubenswrapper[4956]: I1126 17:05:59.385754 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerStarted","Data":"d9b4b5c1518e0cd85b57b43440edbaec7c8a1425657238162c9ee417f428b81a"} Nov 26 17:05:59 crc kubenswrapper[4956]: I1126 17:05:59.387792 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" event={"ID":"495cf683-8b21-4b50-9207-b7727070c747","Type":"ContainerDied","Data":"adc45d9de082b37936a929ad07ba121f8e63ccb7fb53d51fc78dd3f80548f002"} Nov 26 17:05:59 crc kubenswrapper[4956]: I1126 17:05:59.387814 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="adc45d9de082b37936a929ad07ba121f8e63ccb7fb53d51fc78dd3f80548f002" Nov 26 17:05:59 crc kubenswrapper[4956]: I1126 17:05:59.387905 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq" Nov 26 17:05:59 crc kubenswrapper[4956]: I1126 17:05:59.406279 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rwggh" podStartSLOduration=1.804249277 podStartE2EDuration="5.406255404s" podCreationTimestamp="2025-11-26 17:05:54 +0000 UTC" firstStartedPulling="2025-11-26 17:05:55.352172443 +0000 UTC m=+801.048132995" lastFinishedPulling="2025-11-26 17:05:58.95417857 +0000 UTC m=+804.650139122" observedRunningTime="2025-11-26 17:05:59.402987619 +0000 UTC m=+805.098948181" watchObservedRunningTime="2025-11-26 17:05:59.406255404 +0000 UTC m=+805.102215956" Nov 26 17:06:04 crc kubenswrapper[4956]: I1126 17:06:04.371953 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:06:04 crc kubenswrapper[4956]: I1126 17:06:04.372299 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:06:04 crc kubenswrapper[4956]: I1126 17:06:04.432592 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:06:04 crc kubenswrapper[4956]: I1126 17:06:04.476144 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.360951 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99"] Nov 26 17:06:05 crc kubenswrapper[4956]: E1126 17:06:05.362059 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="util" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.362168 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="util" Nov 26 17:06:05 crc kubenswrapper[4956]: E1126 17:06:05.362258 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="pull" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.362336 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="pull" Nov 26 17:06:05 crc kubenswrapper[4956]: E1126 17:06:05.362421 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="extract" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.362507 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="extract" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.362742 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="495cf683-8b21-4b50-9207-b7727070c747" containerName="extract" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.363711 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.365975 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.366367 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-cjzx8" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.375071 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99"] Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.453598 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/063794be-3564-44a8-8613-a966e3ebd953-webhook-cert\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.453752 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/063794be-3564-44a8-8613-a966e3ebd953-apiservice-cert\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.453907 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qf59l\" (UniqueName: \"kubernetes.io/projected/063794be-3564-44a8-8613-a966e3ebd953-kube-api-access-qf59l\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.554469 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qf59l\" (UniqueName: \"kubernetes.io/projected/063794be-3564-44a8-8613-a966e3ebd953-kube-api-access-qf59l\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.554536 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/063794be-3564-44a8-8613-a966e3ebd953-webhook-cert\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.554578 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/063794be-3564-44a8-8613-a966e3ebd953-apiservice-cert\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.562234 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/063794be-3564-44a8-8613-a966e3ebd953-webhook-cert\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.562270 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/063794be-3564-44a8-8613-a966e3ebd953-apiservice-cert\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.582840 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qf59l\" (UniqueName: \"kubernetes.io/projected/063794be-3564-44a8-8613-a966e3ebd953-kube-api-access-qf59l\") pod \"infra-operator-controller-manager-5c4ffc67cb-dnf99\" (UID: \"063794be-3564-44a8-8613-a966e3ebd953\") " pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:05 crc kubenswrapper[4956]: I1126 17:06:05.683609 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:06 crc kubenswrapper[4956]: I1126 17:06:06.809557 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99"] Nov 26 17:06:07 crc kubenswrapper[4956]: I1126 17:06:07.437235 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerStarted","Data":"82d42d3089486555939beaab7060cf8b38daed8c8ce90c77d666757bbc16c25c"} Nov 26 17:06:07 crc kubenswrapper[4956]: I1126 17:06:07.447932 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rwggh"] Nov 26 17:06:07 crc kubenswrapper[4956]: I1126 17:06:07.448220 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rwggh" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="registry-server" containerID="cri-o://d9b4b5c1518e0cd85b57b43440edbaec7c8a1425657238162c9ee417f428b81a" gracePeriod=2 Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.450579 4956 generic.go:334] "Generic (PLEG): container finished" podID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerID="d9b4b5c1518e0cd85b57b43440edbaec7c8a1425657238162c9ee417f428b81a" exitCode=0 Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.450774 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerDied","Data":"d9b4b5c1518e0cd85b57b43440edbaec7c8a1425657238162c9ee417f428b81a"} Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.897472 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.901837 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-catalog-content\") pod \"0b59c522-cd4b-4615-b579-9a5574cc72d4\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.902028 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q98ms\" (UniqueName: \"kubernetes.io/projected/0b59c522-cd4b-4615-b579-9a5574cc72d4-kube-api-access-q98ms\") pod \"0b59c522-cd4b-4615-b579-9a5574cc72d4\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.902074 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-utilities\") pod \"0b59c522-cd4b-4615-b579-9a5574cc72d4\" (UID: \"0b59c522-cd4b-4615-b579-9a5574cc72d4\") " Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.903897 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-utilities" (OuterVolumeSpecName: "utilities") pod "0b59c522-cd4b-4615-b579-9a5574cc72d4" (UID: "0b59c522-cd4b-4615-b579-9a5574cc72d4"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.933200 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b59c522-cd4b-4615-b579-9a5574cc72d4-kube-api-access-q98ms" (OuterVolumeSpecName: "kube-api-access-q98ms") pod "0b59c522-cd4b-4615-b579-9a5574cc72d4" (UID: "0b59c522-cd4b-4615-b579-9a5574cc72d4"). InnerVolumeSpecName "kube-api-access-q98ms". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:06:08 crc kubenswrapper[4956]: I1126 17:06:08.963154 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0b59c522-cd4b-4615-b579-9a5574cc72d4" (UID: "0b59c522-cd4b-4615-b579-9a5574cc72d4"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.004900 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.005087 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0b59c522-cd4b-4615-b579-9a5574cc72d4-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.005120 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q98ms\" (UniqueName: \"kubernetes.io/projected/0b59c522-cd4b-4615-b579-9a5574cc72d4-kube-api-access-q98ms\") on node \"crc\" DevicePath \"\"" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.462933 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rwggh" event={"ID":"0b59c522-cd4b-4615-b579-9a5574cc72d4","Type":"ContainerDied","Data":"b29238e51d7e6a8787dca95a05d61b88a4978be5148e88956320095365e757b8"} Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.463007 4956 scope.go:117] "RemoveContainer" containerID="d9b4b5c1518e0cd85b57b43440edbaec7c8a1425657238162c9ee417f428b81a" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.463054 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rwggh" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.465171 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerStarted","Data":"9ed1d2310d68aeae6b94618f4b7e1f11aff32d6885c3549ffaddd6bc9b04e23a"} Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.488622 4956 scope.go:117] "RemoveContainer" containerID="d1e16f6d33694dd4cb6a8be8ec12fbb14f6ca44980d487fd4c04338bafcbb701" Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.489395 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rwggh"] Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.496885 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rwggh"] Nov 26 17:06:09 crc kubenswrapper[4956]: I1126 17:06:09.510078 4956 scope.go:117] "RemoveContainer" containerID="95ed27c11c47f43a0c7e6887d93ed38f50d639fffc35a53b0c7be109c23bec0f" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.386451 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 26 17:06:10 crc kubenswrapper[4956]: E1126 17:06:10.386735 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="extract-content" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.386754 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="extract-content" Nov 26 17:06:10 crc kubenswrapper[4956]: E1126 17:06:10.386773 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="extract-utilities" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.386780 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="extract-utilities" Nov 26 17:06:10 crc kubenswrapper[4956]: E1126 17:06:10.386790 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="registry-server" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.386796 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="registry-server" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.388219 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" containerName="registry-server" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.389031 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.391077 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.392423 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.392747 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-f55gs" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.392450 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.395631 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.400991 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.430036 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.431347 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.444644 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.452731 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.454198 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.458126 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.531769 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-kolla-config\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.531849 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/13129b6d-631d-405b-8201-df79314c929c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.531903 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7bbmk\" (UniqueName: \"kubernetes.io/projected/13129b6d-631d-405b-8201-df79314c929c-kube-api-access-7bbmk\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.531933 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.531961 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-config-data-default\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.531982 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633533 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-operator-scripts\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633630 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633680 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-kolla-config\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633711 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-config-data-default\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633735 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-config-data-default\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633759 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633793 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633828 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-kolla-config\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633855 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-kolla-config\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633938 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.633996 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634026 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634087 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6fc57249-9b8b-48d0-891b-36181a401d7a-config-data-generated\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634119 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/13129b6d-631d-405b-8201-df79314c929c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634185 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kj5xv\" (UniqueName: \"kubernetes.io/projected/6fc57249-9b8b-48d0-891b-36181a401d7a-kube-api-access-kj5xv\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634217 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7bbmk\" (UniqueName: \"kubernetes.io/projected/13129b6d-631d-405b-8201-df79314c929c-kube-api-access-7bbmk\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634265 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f7mn5\" (UniqueName: \"kubernetes.io/projected/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-kube-api-access-f7mn5\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634376 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-config-data-default\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.634983 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/13129b6d-631d-405b-8201-df79314c929c-config-data-generated\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.635238 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") device mount path \"/mnt/openstack/pv01\"" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.636163 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-operator-scripts\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.637049 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-config-data-default\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.648495 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/13129b6d-631d-405b-8201-df79314c929c-kolla-config\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.658057 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage01-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage01-crc\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.658666 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7bbmk\" (UniqueName: \"kubernetes.io/projected/13129b6d-631d-405b-8201-df79314c929c-kube-api-access-7bbmk\") pod \"openstack-galera-0\" (UID: \"13129b6d-631d-405b-8201-df79314c929c\") " pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.717697 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.736965 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-kolla-config\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737050 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737081 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737110 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737138 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6fc57249-9b8b-48d0-891b-36181a401d7a-config-data-generated\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737176 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kj5xv\" (UniqueName: \"kubernetes.io/projected/6fc57249-9b8b-48d0-891b-36181a401d7a-kube-api-access-kj5xv\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737203 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f7mn5\" (UniqueName: \"kubernetes.io/projected/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-kube-api-access-f7mn5\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737232 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-config-data-default\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737257 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-operator-scripts\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737300 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-kolla-config\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737331 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-config-data-default\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737357 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737475 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") device mount path \"/mnt/openstack/pv14\"" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737559 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") device mount path \"/mnt/openstack/pv05\"" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.737642 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-config-data-generated\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.738513 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-config-data-default\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.738589 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-kolla-config\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.738834 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-config-data-default\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.738888 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/6fc57249-9b8b-48d0-891b-36181a401d7a-config-data-generated\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.739325 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-operator-scripts\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.739505 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6fc57249-9b8b-48d0-891b-36181a401d7a-operator-scripts\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.739768 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-kolla-config\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.757131 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.765891 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kj5xv\" (UniqueName: \"kubernetes.io/projected/6fc57249-9b8b-48d0-891b-36181a401d7a-kube-api-access-kj5xv\") pod \"openstack-galera-1\" (UID: \"6fc57249-9b8b-48d0-891b-36181a401d7a\") " pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.766032 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f7mn5\" (UniqueName: \"kubernetes.io/projected/e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa-kube-api-access-f7mn5\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.773737 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage14-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage14-crc\") pod \"openstack-galera-2\" (UID: \"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa\") " pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:10 crc kubenswrapper[4956]: I1126 17:06:10.779988 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:11 crc kubenswrapper[4956]: I1126 17:06:11.004321 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b59c522-cd4b-4615-b579-9a5574cc72d4" path="/var/lib/kubelet/pods/0b59c522-cd4b-4615-b579-9a5574cc72d4/volumes" Nov 26 17:06:11 crc kubenswrapper[4956]: I1126 17:06:11.055687 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:11 crc kubenswrapper[4956]: I1126 17:06:11.482772 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-0"] Nov 26 17:06:11 crc kubenswrapper[4956]: I1126 17:06:11.519834 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"13129b6d-631d-405b-8201-df79314c929c","Type":"ContainerStarted","Data":"cd1864ea6ede9ba8a33e06abab63463056181972fbef4d2c904cbaf38fa66d7f"} Nov 26 17:06:11 crc kubenswrapper[4956]: I1126 17:06:11.766504 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-2"] Nov 26 17:06:11 crc kubenswrapper[4956]: W1126 17:06:11.775054 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode7deb4ae_2c0e_41e3_9b94_b02426fdbaaa.slice/crio-8cd3aea4e14179748ac566b3da37801942d7e4a25b854df9a22f9fe0e99bee7c WatchSource:0}: Error finding container 8cd3aea4e14179748ac566b3da37801942d7e4a25b854df9a22f9fe0e99bee7c: Status 404 returned error can't find the container with id 8cd3aea4e14179748ac566b3da37801942d7e4a25b854df9a22f9fe0e99bee7c Nov 26 17:06:11 crc kubenswrapper[4956]: I1126 17:06:11.839908 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstack-galera-1"] Nov 26 17:06:11 crc kubenswrapper[4956]: W1126 17:06:11.853029 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6fc57249_9b8b_48d0_891b_36181a401d7a.slice/crio-9e6e64d54bbd62779863e422b1c710e896d9b774b44a8b5dffab9c2cdaa00b8a WatchSource:0}: Error finding container 9e6e64d54bbd62779863e422b1c710e896d9b774b44a8b5dffab9c2cdaa00b8a: Status 404 returned error can't find the container with id 9e6e64d54bbd62779863e422b1c710e896d9b774b44a8b5dffab9c2cdaa00b8a Nov 26 17:06:12 crc kubenswrapper[4956]: I1126 17:06:12.536046 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa","Type":"ContainerStarted","Data":"8cd3aea4e14179748ac566b3da37801942d7e4a25b854df9a22f9fe0e99bee7c"} Nov 26 17:06:12 crc kubenswrapper[4956]: I1126 17:06:12.537430 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"6fc57249-9b8b-48d0-891b-36181a401d7a","Type":"ContainerStarted","Data":"9e6e64d54bbd62779863e422b1c710e896d9b774b44a8b5dffab9c2cdaa00b8a"} Nov 26 17:06:13 crc kubenswrapper[4956]: I1126 17:06:13.546788 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerStarted","Data":"5b6e13507ab05b23adebbf09a3e84bd94ebdcaa82b6f378a946656229e7c5b67"} Nov 26 17:06:13 crc kubenswrapper[4956]: I1126 17:06:13.547568 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:13 crc kubenswrapper[4956]: I1126 17:06:13.569288 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podStartSLOduration=2.947268163 podStartE2EDuration="8.569261855s" podCreationTimestamp="2025-11-26 17:06:05 +0000 UTC" firstStartedPulling="2025-11-26 17:06:06.825766409 +0000 UTC m=+812.521726961" lastFinishedPulling="2025-11-26 17:06:12.447760101 +0000 UTC m=+818.143720653" observedRunningTime="2025-11-26 17:06:13.565467185 +0000 UTC m=+819.261427737" watchObservedRunningTime="2025-11-26 17:06:13.569261855 +0000 UTC m=+819.265222407" Nov 26 17:06:14 crc kubenswrapper[4956]: I1126 17:06:14.560338 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:06:16 crc kubenswrapper[4956]: I1126 17:06:16.895615 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 26 17:06:16 crc kubenswrapper[4956]: I1126 17:06:16.896785 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:16 crc kubenswrapper[4956]: I1126 17:06:16.902442 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 26 17:06:16 crc kubenswrapper[4956]: I1126 17:06:16.902777 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-8t9gl" Nov 26 17:06:16 crc kubenswrapper[4956]: I1126 17:06:16.905409 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.064708 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/199646a6-766c-4047-bced-5bada18d54bf-config-data\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.064774 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/199646a6-766c-4047-bced-5bada18d54bf-kolla-config\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.064802 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v5fm5\" (UniqueName: \"kubernetes.io/projected/199646a6-766c-4047-bced-5bada18d54bf-kube-api-access-v5fm5\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.166340 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/199646a6-766c-4047-bced-5bada18d54bf-config-data\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.166410 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/199646a6-766c-4047-bced-5bada18d54bf-kolla-config\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.166437 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v5fm5\" (UniqueName: \"kubernetes.io/projected/199646a6-766c-4047-bced-5bada18d54bf-kube-api-access-v5fm5\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.167340 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/199646a6-766c-4047-bced-5bada18d54bf-config-data\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.168630 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/199646a6-766c-4047-bced-5bada18d54bf-kolla-config\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.186769 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v5fm5\" (UniqueName: \"kubernetes.io/projected/199646a6-766c-4047-bced-5bada18d54bf-kube-api-access-v5fm5\") pod \"memcached-0\" (UID: \"199646a6-766c-4047-bced-5bada18d54bf\") " pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:17 crc kubenswrapper[4956]: I1126 17:06:17.223786 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.674412 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-vgxbf"] Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.676083 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.698951 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgxbf"] Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.812925 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5102435b-0e33-4be2-a344-df54bd02afa6-catalog-content\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.813000 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p75zt\" (UniqueName: \"kubernetes.io/projected/5102435b-0e33-4be2-a344-df54bd02afa6-kube-api-access-p75zt\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.813026 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5102435b-0e33-4be2-a344-df54bd02afa6-utilities\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.914194 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p75zt\" (UniqueName: \"kubernetes.io/projected/5102435b-0e33-4be2-a344-df54bd02afa6-kube-api-access-p75zt\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.914290 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5102435b-0e33-4be2-a344-df54bd02afa6-utilities\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.914395 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5102435b-0e33-4be2-a344-df54bd02afa6-catalog-content\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.914843 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5102435b-0e33-4be2-a344-df54bd02afa6-utilities\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.915023 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5102435b-0e33-4be2-a344-df54bd02afa6-catalog-content\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.950898 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p75zt\" (UniqueName: \"kubernetes.io/projected/5102435b-0e33-4be2-a344-df54bd02afa6-kube-api-access-p75zt\") pod \"community-operators-vgxbf\" (UID: \"5102435b-0e33-4be2-a344-df54bd02afa6\") " pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:19 crc kubenswrapper[4956]: I1126 17:06:19.997462 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.455007 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-2nksr"] Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.456325 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.463308 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-6f5m5" Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.475134 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-2nksr"] Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.625272 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzgg7\" (UniqueName: \"kubernetes.io/projected/fc6c140a-8045-4034-b13c-03972336f0de-kube-api-access-jzgg7\") pod \"rabbitmq-cluster-operator-index-2nksr\" (UID: \"fc6c140a-8045-4034-b13c-03972336f0de\") " pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.727284 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzgg7\" (UniqueName: \"kubernetes.io/projected/fc6c140a-8045-4034-b13c-03972336f0de-kube-api-access-jzgg7\") pod \"rabbitmq-cluster-operator-index-2nksr\" (UID: \"fc6c140a-8045-4034-b13c-03972336f0de\") " pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.751529 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzgg7\" (UniqueName: \"kubernetes.io/projected/fc6c140a-8045-4034-b13c-03972336f0de-kube-api-access-jzgg7\") pod \"rabbitmq-cluster-operator-index-2nksr\" (UID: \"fc6c140a-8045-4034-b13c-03972336f0de\") " pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:06:20 crc kubenswrapper[4956]: I1126 17:06:20.778853 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.292492 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-index-2nksr"] Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.329295 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/memcached-0"] Nov 26 17:06:22 crc kubenswrapper[4956]: W1126 17:06:22.342069 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod199646a6_766c_4047_bced_5bada18d54bf.slice/crio-9951b207d2acb49613b69f14626ec52d64c815f3dfd19945a20fd914a3abe845 WatchSource:0}: Error finding container 9951b207d2acb49613b69f14626ec52d64c815f3dfd19945a20fd914a3abe845: Status 404 returned error can't find the container with id 9951b207d2acb49613b69f14626ec52d64c815f3dfd19945a20fd914a3abe845 Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.595375 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgxbf"] Nov 26 17:06:22 crc kubenswrapper[4956]: W1126 17:06:22.598604 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5102435b_0e33_4be2_a344_df54bd02afa6.slice/crio-880bed8e7df6505045237c304fea89cd59a194f3cae72617faf86d2e894615ce WatchSource:0}: Error finding container 880bed8e7df6505045237c304fea89cd59a194f3cae72617faf86d2e894615ce: Status 404 returned error can't find the container with id 880bed8e7df6505045237c304fea89cd59a194f3cae72617faf86d2e894615ce Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.612126 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgxbf" event={"ID":"5102435b-0e33-4be2-a344-df54bd02afa6","Type":"ContainerStarted","Data":"880bed8e7df6505045237c304fea89cd59a194f3cae72617faf86d2e894615ce"} Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.613886 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" event={"ID":"fc6c140a-8045-4034-b13c-03972336f0de","Type":"ContainerStarted","Data":"a1e77e848e771d63ae8c79c46df862c74f7a504f2eb1a1223024ceac549ec7eb"} Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.615843 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa","Type":"ContainerStarted","Data":"ef67867e816c4fde28da13f7f93a70d070aae24cdf9af74ae0997be2bb59c7ee"} Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.618631 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"199646a6-766c-4047-bced-5bada18d54bf","Type":"ContainerStarted","Data":"9951b207d2acb49613b69f14626ec52d64c815f3dfd19945a20fd914a3abe845"} Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.621188 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"6fc57249-9b8b-48d0-891b-36181a401d7a","Type":"ContainerStarted","Data":"50c0976b6c874d3a9fb4096e9468a22d4ad64f9869adc80f0000926ee89d45ff"} Nov 26 17:06:22 crc kubenswrapper[4956]: I1126 17:06:22.623643 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"13129b6d-631d-405b-8201-df79314c929c","Type":"ContainerStarted","Data":"f90a8caae198f3a38730623f1087475252a97f2b8b355cfdf26a2a5c39941a10"} Nov 26 17:06:23 crc kubenswrapper[4956]: I1126 17:06:23.635670 4956 generic.go:334] "Generic (PLEG): container finished" podID="5102435b-0e33-4be2-a344-df54bd02afa6" containerID="73c8674eeb62f35fd1557a4601b54b30bc2ac91131c38d742ceddf16a73435f1" exitCode=0 Nov 26 17:06:23 crc kubenswrapper[4956]: I1126 17:06:23.635831 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgxbf" event={"ID":"5102435b-0e33-4be2-a344-df54bd02afa6","Type":"ContainerDied","Data":"73c8674eeb62f35fd1557a4601b54b30bc2ac91131c38d742ceddf16a73435f1"} Nov 26 17:06:32 crc kubenswrapper[4956]: I1126 17:06:32.701557 4956 generic.go:334] "Generic (PLEG): container finished" podID="e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa" containerID="ef67867e816c4fde28da13f7f93a70d070aae24cdf9af74ae0997be2bb59c7ee" exitCode=0 Nov 26 17:06:32 crc kubenswrapper[4956]: I1126 17:06:32.701693 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa","Type":"ContainerDied","Data":"ef67867e816c4fde28da13f7f93a70d070aae24cdf9af74ae0997be2bb59c7ee"} Nov 26 17:06:33 crc kubenswrapper[4956]: I1126 17:06:33.710552 4956 generic.go:334] "Generic (PLEG): container finished" podID="13129b6d-631d-405b-8201-df79314c929c" containerID="f90a8caae198f3a38730623f1087475252a97f2b8b355cfdf26a2a5c39941a10" exitCode=0 Nov 26 17:06:33 crc kubenswrapper[4956]: I1126 17:06:33.710608 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"13129b6d-631d-405b-8201-df79314c929c","Type":"ContainerDied","Data":"f90a8caae198f3a38730623f1087475252a97f2b8b355cfdf26a2a5c39941a10"} Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.265945 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h69hg"] Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.268228 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.279152 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h69hg"] Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.456022 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mtk66\" (UniqueName: \"kubernetes.io/projected/c603ef85-6178-4c88-aab1-e795dedb1e3d-kube-api-access-mtk66\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.456182 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-utilities\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.456345 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-catalog-content\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.558246 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mtk66\" (UniqueName: \"kubernetes.io/projected/c603ef85-6178-4c88-aab1-e795dedb1e3d-kube-api-access-mtk66\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.558346 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-utilities\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.558466 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-catalog-content\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.618437 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-utilities\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.618558 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-catalog-content\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.628321 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mtk66\" (UniqueName: \"kubernetes.io/projected/c603ef85-6178-4c88-aab1-e795dedb1e3d-kube-api-access-mtk66\") pod \"redhat-operators-h69hg\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.718398 4956 generic.go:334] "Generic (PLEG): container finished" podID="6fc57249-9b8b-48d0-891b-36181a401d7a" containerID="50c0976b6c874d3a9fb4096e9468a22d4ad64f9869adc80f0000926ee89d45ff" exitCode=0 Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.718476 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"6fc57249-9b8b-48d0-891b-36181a401d7a","Type":"ContainerDied","Data":"50c0976b6c874d3a9fb4096e9468a22d4ad64f9869adc80f0000926ee89d45ff"} Nov 26 17:06:34 crc kubenswrapper[4956]: I1126 17:06:34.919776 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:35 crc kubenswrapper[4956]: E1126 17:06:35.695541 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2" Nov 26 17:06:35 crc kubenswrapper[4956]: E1126 17:06:35.696243 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:memcached,Image:quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2,Command:[/usr/bin/dumb-init -- /usr/local/bin/kolla_start],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:memcached,HostPort:0,ContainerPort:11211,Protocol:TCP,HostIP:,},ContainerPort{Name:memcached-tls,HostPort:0,ContainerPort:11212,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},EnvVar{Name:POD_IPS,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIPs,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:CONFIG_HASH,Value:nd6h9dh5dfh5fhdch94h54fh59fh74h679h5dbh665h67h68hfch54fh5ffhchcfh5f4hbbh679h5cch654h57fh644h59ch66fh586h596h6ch54cq,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/src,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kolla-config,ReadOnly:true,MountPath:/var/lib/kolla/config_files,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-v5fm5,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:3,TimeoutSeconds:5,PeriodSeconds:3,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:nil,TCPSocket:&TCPSocketAction{Port:{0 11211 },Host:,},GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:5,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*42457,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:*42457,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod memcached-0_glance-kuttl-tests(199646a6-766c-4047-bced-5bada18d54bf): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 17:06:35 crc kubenswrapper[4956]: E1126 17:06:35.704179 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="glance-kuttl-tests/memcached-0" podUID="199646a6-766c-4047-bced-5bada18d54bf" Nov 26 17:06:35 crc kubenswrapper[4956]: E1126 17:06:35.730586 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"memcached\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-memcached@sha256:36a0fb31978aee0ded2483de311631e64a644d0b0685b5b055f65ede7eb8e8a2\\\"\"" pod="glance-kuttl-tests/memcached-0" podUID="199646a6-766c-4047-bced-5bada18d54bf" Nov 26 17:06:42 crc kubenswrapper[4956]: I1126 17:06:42.214886 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h69hg"] Nov 26 17:06:42 crc kubenswrapper[4956]: I1126 17:06:42.811141 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerStarted","Data":"071082a395d99cdbea93b32e96d927d615b59c2828033693c191ad11f2f30c65"} Nov 26 17:06:45 crc kubenswrapper[4956]: E1126 17:06:45.072982 4956 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest" Nov 26 17:06:45 crc kubenswrapper[4956]: E1126 17:06:45.073633 4956 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:registry-server,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest,Command:[],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:grpc,HostPort:0,ContainerPort:50051,Protocol:TCP,HostIP:,},},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{52428800 0} {} 50Mi BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-jzgg7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:10,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:*false,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:&Probe{ProbeHandler:ProbeHandler{Exec:&ExecAction{Command:[grpc_health_probe -addr=:50051],},HTTPGet:nil,TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:0,TimeoutSeconds:5,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:10,TerminationGracePeriodSeconds:nil,},ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-index-2nksr_openstack-operators(fc6c140a-8045-4034-b13c-03972336f0de): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Nov 26 17:06:45 crc kubenswrapper[4956]: E1126 17:06:45.074986 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" podUID="fc6c140a-8045-4034-b13c-03972336f0de" Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.840305 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-0" event={"ID":"13129b6d-631d-405b-8201-df79314c929c","Type":"ContainerStarted","Data":"953aee45766df04d4456dde48c06df8f4953cc55d4214b2a1e790a4caee0579d"} Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.844599 4956 generic.go:334] "Generic (PLEG): container finished" podID="5102435b-0e33-4be2-a344-df54bd02afa6" containerID="f8b0392093e8ed6cfb35652c5632fc0e014697e5fc5cac3d61cbe649e9525d90" exitCode=0 Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.844648 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgxbf" event={"ID":"5102435b-0e33-4be2-a344-df54bd02afa6","Type":"ContainerDied","Data":"f8b0392093e8ed6cfb35652c5632fc0e014697e5fc5cac3d61cbe649e9525d90"} Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.847400 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-2" event={"ID":"e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa","Type":"ContainerStarted","Data":"2c3839623852f4f6291db995ad50694891d881597fdd0e7a3a92e7ee1fd2b106"} Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.850345 4956 generic.go:334] "Generic (PLEG): container finished" podID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerID="f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6" exitCode=0 Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.850406 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerDied","Data":"f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6"} Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.863701 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-0" podStartSLOduration=26.311031858 podStartE2EDuration="36.86367627s" podCreationTimestamp="2025-11-26 17:06:09 +0000 UTC" firstStartedPulling="2025-11-26 17:06:11.494344263 +0000 UTC m=+817.190304815" lastFinishedPulling="2025-11-26 17:06:22.046988675 +0000 UTC m=+827.742949227" observedRunningTime="2025-11-26 17:06:45.860932671 +0000 UTC m=+851.556893243" watchObservedRunningTime="2025-11-26 17:06:45.86367627 +0000 UTC m=+851.559636832" Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.864438 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstack-galera-1" event={"ID":"6fc57249-9b8b-48d0-891b-36181a401d7a","Type":"ContainerStarted","Data":"cd5fab292852563828ae9763b154c337ef12d9ac4b1d13a5379cd1bc0c375791"} Nov 26 17:06:45 crc kubenswrapper[4956]: E1126 17:06:45.865811 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"registry-server\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator-index:latest\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" podUID="fc6c140a-8045-4034-b13c-03972336f0de" Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.892592 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-2" podStartSLOduration=26.633955342 podStartE2EDuration="36.892563706s" podCreationTimestamp="2025-11-26 17:06:09 +0000 UTC" firstStartedPulling="2025-11-26 17:06:11.778550438 +0000 UTC m=+817.474510980" lastFinishedPulling="2025-11-26 17:06:22.037158802 +0000 UTC m=+827.733119344" observedRunningTime="2025-11-26 17:06:45.887437068 +0000 UTC m=+851.583397630" watchObservedRunningTime="2025-11-26 17:06:45.892563706 +0000 UTC m=+851.588524268" Nov 26 17:06:45 crc kubenswrapper[4956]: I1126 17:06:45.954652 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstack-galera-1" podStartSLOduration=26.782085703 podStartE2EDuration="36.95458488s" podCreationTimestamp="2025-11-26 17:06:09 +0000 UTC" firstStartedPulling="2025-11-26 17:06:11.865480428 +0000 UTC m=+817.561440980" lastFinishedPulling="2025-11-26 17:06:22.037979605 +0000 UTC m=+827.733940157" observedRunningTime="2025-11-26 17:06:45.952454424 +0000 UTC m=+851.648414976" watchObservedRunningTime="2025-11-26 17:06:45.95458488 +0000 UTC m=+851.650545482" Nov 26 17:06:47 crc kubenswrapper[4956]: I1126 17:06:47.909298 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerStarted","Data":"f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e"} Nov 26 17:06:47 crc kubenswrapper[4956]: I1126 17:06:47.925949 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-vgxbf" event={"ID":"5102435b-0e33-4be2-a344-df54bd02afa6","Type":"ContainerStarted","Data":"50d1d065173fb2693b062d617bbd4a46eae9b2025e42213add29f5b394476be4"} Nov 26 17:06:47 crc kubenswrapper[4956]: I1126 17:06:47.957663 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-vgxbf" podStartSLOduration=11.656497703 podStartE2EDuration="28.957639564s" podCreationTimestamp="2025-11-26 17:06:19 +0000 UTC" firstStartedPulling="2025-11-26 17:06:30.083791413 +0000 UTC m=+835.779751965" lastFinishedPulling="2025-11-26 17:06:47.384933264 +0000 UTC m=+853.080893826" observedRunningTime="2025-11-26 17:06:47.952303565 +0000 UTC m=+853.648264117" watchObservedRunningTime="2025-11-26 17:06:47.957639564 +0000 UTC m=+853.653600116" Nov 26 17:06:48 crc kubenswrapper[4956]: I1126 17:06:48.935370 4956 generic.go:334] "Generic (PLEG): container finished" podID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerID="f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e" exitCode=0 Nov 26 17:06:48 crc kubenswrapper[4956]: I1126 17:06:48.935474 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerDied","Data":"f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e"} Nov 26 17:06:49 crc kubenswrapper[4956]: I1126 17:06:49.944171 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/memcached-0" event={"ID":"199646a6-766c-4047-bced-5bada18d54bf","Type":"ContainerStarted","Data":"398448cefc177df6df293352bfa9a4964330689eb8fde11f06ea83f3b7e387f4"} Nov 26 17:06:49 crc kubenswrapper[4956]: I1126 17:06:49.944984 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:49 crc kubenswrapper[4956]: I1126 17:06:49.964589 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/memcached-0" podStartSLOduration=7.384022283 podStartE2EDuration="33.964557438s" podCreationTimestamp="2025-11-26 17:06:16 +0000 UTC" firstStartedPulling="2025-11-26 17:06:22.345266487 +0000 UTC m=+828.041227039" lastFinishedPulling="2025-11-26 17:06:48.925801642 +0000 UTC m=+854.621762194" observedRunningTime="2025-11-26 17:06:49.959809751 +0000 UTC m=+855.655770303" watchObservedRunningTime="2025-11-26 17:06:49.964557438 +0000 UTC m=+855.660518000" Nov 26 17:06:49 crc kubenswrapper[4956]: I1126 17:06:49.998592 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:49 crc kubenswrapper[4956]: I1126 17:06:49.998661 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:50 crc kubenswrapper[4956]: I1126 17:06:50.049509 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:06:50 crc kubenswrapper[4956]: I1126 17:06:50.718084 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:50 crc kubenswrapper[4956]: I1126 17:06:50.718675 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:06:50 crc kubenswrapper[4956]: I1126 17:06:50.780236 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:50 crc kubenswrapper[4956]: I1126 17:06:50.780307 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:06:50 crc kubenswrapper[4956]: I1126 17:06:50.953346 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerStarted","Data":"a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f"} Nov 26 17:06:51 crc kubenswrapper[4956]: I1126 17:06:51.056418 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:51 crc kubenswrapper[4956]: I1126 17:06:51.056468 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:51 crc kubenswrapper[4956]: I1126 17:06:51.982688 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h69hg" podStartSLOduration=14.245633928 podStartE2EDuration="17.982661993s" podCreationTimestamp="2025-11-26 17:06:34 +0000 UTC" firstStartedPulling="2025-11-26 17:06:45.855089414 +0000 UTC m=+851.551049966" lastFinishedPulling="2025-11-26 17:06:49.592117479 +0000 UTC m=+855.288078031" observedRunningTime="2025-11-26 17:06:51.981201562 +0000 UTC m=+857.677162134" watchObservedRunningTime="2025-11-26 17:06:51.982661993 +0000 UTC m=+857.678622555" Nov 26 17:06:54 crc kubenswrapper[4956]: I1126 17:06:54.920953 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:54 crc kubenswrapper[4956]: I1126 17:06:54.922550 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:06:55 crc kubenswrapper[4956]: I1126 17:06:55.573278 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:55 crc kubenswrapper[4956]: I1126 17:06:55.660404 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-2" Nov 26 17:06:55 crc kubenswrapper[4956]: I1126 17:06:55.970755 4956 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h69hg" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="registry-server" probeResult="failure" output=< Nov 26 17:06:55 crc kubenswrapper[4956]: timeout: failed to connect service ":50051" within 1s Nov 26 17:06:55 crc kubenswrapper[4956]: > Nov 26 17:06:57 crc kubenswrapper[4956]: I1126 17:06:57.224723 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/memcached-0" Nov 26 17:06:58 crc kubenswrapper[4956]: I1126 17:06:58.061595 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" event={"ID":"fc6c140a-8045-4034-b13c-03972336f0de","Type":"ContainerStarted","Data":"6708521cade59f88b649e46f66223d8c9db2e257dd37d1268910fa67a29da784"} Nov 26 17:06:58 crc kubenswrapper[4956]: I1126 17:06:58.078287 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" podStartSLOduration=2.67040808 podStartE2EDuration="38.078260051s" podCreationTimestamp="2025-11-26 17:06:20 +0000 UTC" firstStartedPulling="2025-11-26 17:06:22.307972471 +0000 UTC m=+828.003933023" lastFinishedPulling="2025-11-26 17:06:57.715824432 +0000 UTC m=+863.411784994" observedRunningTime="2025-11-26 17:06:58.075432011 +0000 UTC m=+863.771392583" watchObservedRunningTime="2025-11-26 17:06:58.078260051 +0000 UTC m=+863.774220603" Nov 26 17:07:00 crc kubenswrapper[4956]: I1126 17:07:00.044184 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-vgxbf" Nov 26 17:07:00 crc kubenswrapper[4956]: I1126 17:07:00.779622 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:07:00 crc kubenswrapper[4956]: I1126 17:07:00.779716 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:07:00 crc kubenswrapper[4956]: I1126 17:07:00.814622 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:07:01 crc kubenswrapper[4956]: I1126 17:07:01.188982 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/openstack-galera-2" podUID="e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa" containerName="galera" probeResult="failure" output=< Nov 26 17:07:01 crc kubenswrapper[4956]: wsrep_local_state_comment (Donor/Desynced) differs from Synced Nov 26 17:07:01 crc kubenswrapper[4956]: > Nov 26 17:07:03 crc kubenswrapper[4956]: I1126 17:07:03.270165 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-vgxbf"] Nov 26 17:07:03 crc kubenswrapper[4956]: I1126 17:07:03.845929 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dhxg8"] Nov 26 17:07:03 crc kubenswrapper[4956]: I1126 17:07:03.846329 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-dhxg8" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="registry-server" containerID="cri-o://1398981dbf3f4bdf02ce5fd80fb0738e1d1b5debeb1ee1683390cc84a6dcf374" gracePeriod=2 Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.129925 4956 generic.go:334] "Generic (PLEG): container finished" podID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerID="1398981dbf3f4bdf02ce5fd80fb0738e1d1b5debeb1ee1683390cc84a6dcf374" exitCode=0 Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.130034 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerDied","Data":"1398981dbf3f4bdf02ce5fd80fb0738e1d1b5debeb1ee1683390cc84a6dcf374"} Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.355701 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.537143 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9pbsn\" (UniqueName: \"kubernetes.io/projected/f5bfa256-d54e-4aa1-a63c-7805fb956688-kube-api-access-9pbsn\") pod \"f5bfa256-d54e-4aa1-a63c-7805fb956688\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.537374 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-utilities\") pod \"f5bfa256-d54e-4aa1-a63c-7805fb956688\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.537457 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-catalog-content\") pod \"f5bfa256-d54e-4aa1-a63c-7805fb956688\" (UID: \"f5bfa256-d54e-4aa1-a63c-7805fb956688\") " Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.538524 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-utilities" (OuterVolumeSpecName: "utilities") pod "f5bfa256-d54e-4aa1-a63c-7805fb956688" (UID: "f5bfa256-d54e-4aa1-a63c-7805fb956688"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.556351 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f5bfa256-d54e-4aa1-a63c-7805fb956688-kube-api-access-9pbsn" (OuterVolumeSpecName: "kube-api-access-9pbsn") pod "f5bfa256-d54e-4aa1-a63c-7805fb956688" (UID: "f5bfa256-d54e-4aa1-a63c-7805fb956688"). InnerVolumeSpecName "kube-api-access-9pbsn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.590736 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f5bfa256-d54e-4aa1-a63c-7805fb956688" (UID: "f5bfa256-d54e-4aa1-a63c-7805fb956688"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.639223 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.639486 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f5bfa256-d54e-4aa1-a63c-7805fb956688-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.639565 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9pbsn\" (UniqueName: \"kubernetes.io/projected/f5bfa256-d54e-4aa1-a63c-7805fb956688-kube-api-access-9pbsn\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:04 crc kubenswrapper[4956]: I1126 17:07:04.972288 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.019740 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.139916 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dhxg8" event={"ID":"f5bfa256-d54e-4aa1-a63c-7805fb956688","Type":"ContainerDied","Data":"9674a2da22440107edab1c13ca84e9e65457fa9fa4020b74c937273296a4b37d"} Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.139956 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dhxg8" Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.140000 4956 scope.go:117] "RemoveContainer" containerID="1398981dbf3f4bdf02ce5fd80fb0738e1d1b5debeb1ee1683390cc84a6dcf374" Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.161833 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-dhxg8"] Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.164946 4956 scope.go:117] "RemoveContainer" containerID="551d104c95a65107374d63743e3464cf6c93f4d7b1a067474d38a1de5641d4bd" Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.166203 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-dhxg8"] Nov 26 17:07:05 crc kubenswrapper[4956]: I1126 17:07:05.183585 4956 scope.go:117] "RemoveContainer" containerID="2e3fee9b24fa6d80c28e4b559c97c679a25644f01dc8aa71dc43aab2ef50c19f" Nov 26 17:07:07 crc kubenswrapper[4956]: I1126 17:07:07.004680 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" path="/var/lib/kubelet/pods/f5bfa256-d54e-4aa1-a63c-7805fb956688/volumes" Nov 26 17:07:09 crc kubenswrapper[4956]: I1126 17:07:09.413413 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:07:09 crc kubenswrapper[4956]: I1126 17:07:09.445955 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h69hg"] Nov 26 17:07:09 crc kubenswrapper[4956]: I1126 17:07:09.446270 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h69hg" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="registry-server" containerID="cri-o://a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f" gracePeriod=2 Nov 26 17:07:09 crc kubenswrapper[4956]: I1126 17:07:09.493730 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-1" Nov 26 17:07:09 crc kubenswrapper[4956]: I1126 17:07:09.855721 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.024968 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-utilities\") pod \"c603ef85-6178-4c88-aab1-e795dedb1e3d\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.025170 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-catalog-content\") pod \"c603ef85-6178-4c88-aab1-e795dedb1e3d\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.025232 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mtk66\" (UniqueName: \"kubernetes.io/projected/c603ef85-6178-4c88-aab1-e795dedb1e3d-kube-api-access-mtk66\") pod \"c603ef85-6178-4c88-aab1-e795dedb1e3d\" (UID: \"c603ef85-6178-4c88-aab1-e795dedb1e3d\") " Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.026772 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-utilities" (OuterVolumeSpecName: "utilities") pod "c603ef85-6178-4c88-aab1-e795dedb1e3d" (UID: "c603ef85-6178-4c88-aab1-e795dedb1e3d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.034310 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c603ef85-6178-4c88-aab1-e795dedb1e3d-kube-api-access-mtk66" (OuterVolumeSpecName: "kube-api-access-mtk66") pod "c603ef85-6178-4c88-aab1-e795dedb1e3d" (UID: "c603ef85-6178-4c88-aab1-e795dedb1e3d"). InnerVolumeSpecName "kube-api-access-mtk66". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.128060 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "c603ef85-6178-4c88-aab1-e795dedb1e3d" (UID: "c603ef85-6178-4c88-aab1-e795dedb1e3d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.128826 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.128889 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/c603ef85-6178-4c88-aab1-e795dedb1e3d-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.128902 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mtk66\" (UniqueName: \"kubernetes.io/projected/c603ef85-6178-4c88-aab1-e795dedb1e3d-kube-api-access-mtk66\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.175219 4956 generic.go:334] "Generic (PLEG): container finished" podID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerID="a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f" exitCode=0 Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.175297 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h69hg" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.175300 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerDied","Data":"a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f"} Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.175365 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h69hg" event={"ID":"c603ef85-6178-4c88-aab1-e795dedb1e3d","Type":"ContainerDied","Data":"071082a395d99cdbea93b32e96d927d615b59c2828033693c191ad11f2f30c65"} Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.175394 4956 scope.go:117] "RemoveContainer" containerID="a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.203261 4956 scope.go:117] "RemoveContainer" containerID="f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.205566 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h69hg"] Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.209490 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h69hg"] Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.225129 4956 scope.go:117] "RemoveContainer" containerID="f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.250195 4956 scope.go:117] "RemoveContainer" containerID="a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f" Nov 26 17:07:10 crc kubenswrapper[4956]: E1126 17:07:10.252069 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f\": container with ID starting with a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f not found: ID does not exist" containerID="a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.252105 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f"} err="failed to get container status \"a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f\": rpc error: code = NotFound desc = could not find container \"a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f\": container with ID starting with a09f36675e7be4190f534ad58d2ff954f842802f975b3e1d2327f5c4f20de27f not found: ID does not exist" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.252133 4956 scope.go:117] "RemoveContainer" containerID="f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e" Nov 26 17:07:10 crc kubenswrapper[4956]: E1126 17:07:10.252523 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e\": container with ID starting with f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e not found: ID does not exist" containerID="f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.252592 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e"} err="failed to get container status \"f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e\": rpc error: code = NotFound desc = could not find container \"f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e\": container with ID starting with f0e7c8f858069c1aba4d915d6047739b19c8421946084acda85e131ca8e0c38e not found: ID does not exist" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.252632 4956 scope.go:117] "RemoveContainer" containerID="f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6" Nov 26 17:07:10 crc kubenswrapper[4956]: E1126 17:07:10.252976 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6\": container with ID starting with f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6 not found: ID does not exist" containerID="f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.253012 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6"} err="failed to get container status \"f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6\": rpc error: code = NotFound desc = could not find container \"f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6\": container with ID starting with f518ab8a30f6ec65808973c375601ee85eb78aed3610d2bf6d8f89b25a0addf6 not found: ID does not exist" Nov 26 17:07:10 crc kubenswrapper[4956]: I1126 17:07:10.821787 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/rabbitmq-cluster-operator-index-2nksr" Nov 26 17:07:11 crc kubenswrapper[4956]: I1126 17:07:11.007743 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" path="/var/lib/kubelet/pods/c603ef85-6178-4c88-aab1-e795dedb1e3d/volumes" Nov 26 17:07:11 crc kubenswrapper[4956]: I1126 17:07:11.771078 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:07:11 crc kubenswrapper[4956]: I1126 17:07:11.859426 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/openstack-galera-0" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.114291 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x"] Nov 26 17:07:45 crc kubenswrapper[4956]: E1126 17:07:45.115608 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="extract-utilities" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.116094 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="extract-utilities" Nov 26 17:07:45 crc kubenswrapper[4956]: E1126 17:07:45.116166 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="registry-server" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.116173 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="registry-server" Nov 26 17:07:45 crc kubenswrapper[4956]: E1126 17:07:45.116187 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="extract-utilities" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.116196 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="extract-utilities" Nov 26 17:07:45 crc kubenswrapper[4956]: E1126 17:07:45.116206 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="extract-content" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.116212 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="extract-content" Nov 26 17:07:45 crc kubenswrapper[4956]: E1126 17:07:45.116241 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="registry-server" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.116247 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="registry-server" Nov 26 17:07:45 crc kubenswrapper[4956]: E1126 17:07:45.116259 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="extract-content" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.116265 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="extract-content" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.117967 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="c603ef85-6178-4c88-aab1-e795dedb1e3d" containerName="registry-server" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.118033 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f5bfa256-d54e-4aa1-a63c-7805fb956688" containerName="registry-server" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.119896 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.124704 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j97nd" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.132898 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x"] Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.214767 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.214913 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.214968 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz4m7\" (UniqueName: \"kubernetes.io/projected/73d12891-f813-4273-b9a8-de53a4067e14-kube-api-access-tz4m7\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.316829 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.317249 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz4m7\" (UniqueName: \"kubernetes.io/projected/73d12891-f813-4273-b9a8-de53a4067e14-kube-api-access-tz4m7\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.317376 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.318176 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-util\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.318543 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-bundle\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.339898 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz4m7\" (UniqueName: \"kubernetes.io/projected/73d12891-f813-4273-b9a8-de53a4067e14-kube-api-access-tz4m7\") pod \"9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.458064 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:45 crc kubenswrapper[4956]: I1126 17:07:45.899472 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x"] Nov 26 17:07:46 crc kubenswrapper[4956]: I1126 17:07:46.485454 4956 generic.go:334] "Generic (PLEG): container finished" podID="73d12891-f813-4273-b9a8-de53a4067e14" containerID="584306b3e0d40f1bdf0a8e0a743a2812cae811c07847abac0b7129fbe3715390" exitCode=0 Nov 26 17:07:46 crc kubenswrapper[4956]: I1126 17:07:46.485541 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" event={"ID":"73d12891-f813-4273-b9a8-de53a4067e14","Type":"ContainerDied","Data":"584306b3e0d40f1bdf0a8e0a743a2812cae811c07847abac0b7129fbe3715390"} Nov 26 17:07:46 crc kubenswrapper[4956]: I1126 17:07:46.485892 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" event={"ID":"73d12891-f813-4273-b9a8-de53a4067e14","Type":"ContainerStarted","Data":"f5adaefc3b99f74ab1e1ea0ebc489827ea3adabbc92e332248882a0e61bea625"} Nov 26 17:07:47 crc kubenswrapper[4956]: I1126 17:07:47.494194 4956 generic.go:334] "Generic (PLEG): container finished" podID="73d12891-f813-4273-b9a8-de53a4067e14" containerID="4c8696b2281504cff5fb200ae1344f015e04a64a14c5f91ca4c12e11dbddbf70" exitCode=0 Nov 26 17:07:47 crc kubenswrapper[4956]: I1126 17:07:47.494301 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" event={"ID":"73d12891-f813-4273-b9a8-de53a4067e14","Type":"ContainerDied","Data":"4c8696b2281504cff5fb200ae1344f015e04a64a14c5f91ca4c12e11dbddbf70"} Nov 26 17:07:47 crc kubenswrapper[4956]: E1126 17:07:47.611561 4956 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73d12891_f813_4273_b9a8_de53a4067e14.slice/crio-conmon-4c8696b2281504cff5fb200ae1344f015e04a64a14c5f91ca4c12e11dbddbf70.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod73d12891_f813_4273_b9a8_de53a4067e14.slice/crio-4c8696b2281504cff5fb200ae1344f015e04a64a14c5f91ca4c12e11dbddbf70.scope\": RecentStats: unable to find data in memory cache]" Nov 26 17:07:48 crc kubenswrapper[4956]: I1126 17:07:48.507292 4956 generic.go:334] "Generic (PLEG): container finished" podID="73d12891-f813-4273-b9a8-de53a4067e14" containerID="8d0033488bf4a373ed2604f6b647f23ddfc6414077781d56c96c5a2ca96bbeb4" exitCode=0 Nov 26 17:07:48 crc kubenswrapper[4956]: I1126 17:07:48.507767 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" event={"ID":"73d12891-f813-4273-b9a8-de53a4067e14","Type":"ContainerDied","Data":"8d0033488bf4a373ed2604f6b647f23ddfc6414077781d56c96c5a2ca96bbeb4"} Nov 26 17:07:49 crc kubenswrapper[4956]: I1126 17:07:49.841562 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:49 crc kubenswrapper[4956]: I1126 17:07:49.998960 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz4m7\" (UniqueName: \"kubernetes.io/projected/73d12891-f813-4273-b9a8-de53a4067e14-kube-api-access-tz4m7\") pod \"73d12891-f813-4273-b9a8-de53a4067e14\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " Nov 26 17:07:49 crc kubenswrapper[4956]: I1126 17:07:49.999077 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-util\") pod \"73d12891-f813-4273-b9a8-de53a4067e14\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " Nov 26 17:07:49 crc kubenswrapper[4956]: I1126 17:07:49.999368 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-bundle\") pod \"73d12891-f813-4273-b9a8-de53a4067e14\" (UID: \"73d12891-f813-4273-b9a8-de53a4067e14\") " Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.001142 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-bundle" (OuterVolumeSpecName: "bundle") pod "73d12891-f813-4273-b9a8-de53a4067e14" (UID: "73d12891-f813-4273-b9a8-de53a4067e14"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.023385 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/73d12891-f813-4273-b9a8-de53a4067e14-kube-api-access-tz4m7" (OuterVolumeSpecName: "kube-api-access-tz4m7") pod "73d12891-f813-4273-b9a8-de53a4067e14" (UID: "73d12891-f813-4273-b9a8-de53a4067e14"). InnerVolumeSpecName "kube-api-access-tz4m7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.028713 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-util" (OuterVolumeSpecName: "util") pod "73d12891-f813-4273-b9a8-de53a4067e14" (UID: "73d12891-f813-4273-b9a8-de53a4067e14"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.103263 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.103304 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz4m7\" (UniqueName: \"kubernetes.io/projected/73d12891-f813-4273-b9a8-de53a4067e14-kube-api-access-tz4m7\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.103320 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/73d12891-f813-4273-b9a8-de53a4067e14-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.525764 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" event={"ID":"73d12891-f813-4273-b9a8-de53a4067e14","Type":"ContainerDied","Data":"f5adaefc3b99f74ab1e1ea0ebc489827ea3adabbc92e332248882a0e61bea625"} Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.525821 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f5adaefc3b99f74ab1e1ea0ebc489827ea3adabbc92e332248882a0e61bea625" Nov 26 17:07:50 crc kubenswrapper[4956]: I1126 17:07:50.525903 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.138418 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc"] Nov 26 17:07:55 crc kubenswrapper[4956]: E1126 17:07:55.140850 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="util" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.140970 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="util" Nov 26 17:07:55 crc kubenswrapper[4956]: E1126 17:07:55.141041 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="pull" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.141091 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="pull" Nov 26 17:07:55 crc kubenswrapper[4956]: E1126 17:07:55.141149 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="extract" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.141200 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="extract" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.141448 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="73d12891-f813-4273-b9a8-de53a4067e14" containerName="extract" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.142132 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.149187 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-rlp4r" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.155233 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc"] Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.251014 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bfj8v\" (UniqueName: \"kubernetes.io/projected/25d0fbbb-c310-46d0-ae91-a9804761e8a3-kube-api-access-bfj8v\") pod \"rabbitmq-cluster-operator-779fc9694b-94bjc\" (UID: \"25d0fbbb-c310-46d0-ae91-a9804761e8a3\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.353046 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bfj8v\" (UniqueName: \"kubernetes.io/projected/25d0fbbb-c310-46d0-ae91-a9804761e8a3-kube-api-access-bfj8v\") pod \"rabbitmq-cluster-operator-779fc9694b-94bjc\" (UID: \"25d0fbbb-c310-46d0-ae91-a9804761e8a3\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.376804 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bfj8v\" (UniqueName: \"kubernetes.io/projected/25d0fbbb-c310-46d0-ae91-a9804761e8a3-kube-api-access-bfj8v\") pod \"rabbitmq-cluster-operator-779fc9694b-94bjc\" (UID: \"25d0fbbb-c310-46d0-ae91-a9804761e8a3\") " pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.497679 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" Nov 26 17:07:55 crc kubenswrapper[4956]: I1126 17:07:55.740939 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc"] Nov 26 17:07:56 crc kubenswrapper[4956]: I1126 17:07:56.601740 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" event={"ID":"25d0fbbb-c310-46d0-ae91-a9804761e8a3","Type":"ContainerStarted","Data":"8ed29319462adcea2c02c0948a0843053d4ec151814ac76e77312a10c3aecd99"} Nov 26 17:07:59 crc kubenswrapper[4956]: I1126 17:07:59.675518 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" event={"ID":"25d0fbbb-c310-46d0-ae91-a9804761e8a3","Type":"ContainerStarted","Data":"dfb90541317792b4217df3fad3be8ded2c8630b4550de8f6e64c20b9024f753e"} Nov 26 17:07:59 crc kubenswrapper[4956]: I1126 17:07:59.704538 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" podStartSLOduration=1.175045452 podStartE2EDuration="4.704512399s" podCreationTimestamp="2025-11-26 17:07:55 +0000 UTC" firstStartedPulling="2025-11-26 17:07:55.751554764 +0000 UTC m=+921.447515326" lastFinishedPulling="2025-11-26 17:07:59.281021721 +0000 UTC m=+924.976982273" observedRunningTime="2025-11-26 17:07:59.69719097 +0000 UTC m=+925.393151522" watchObservedRunningTime="2025-11-26 17:07:59.704512399 +0000 UTC m=+925.400472991" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.568189 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.570250 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.572812 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.573242 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.573242 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.573641 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-cjjk9" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.582897 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.632790 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.687885 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.687956 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w86zh\" (UniqueName: \"kubernetes.io/projected/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-kube-api-access-w86zh\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.687999 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pvc-b4484dbd-034d-48f1-a061-680e7edce5be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4484dbd-034d-48f1-a061-680e7edce5be\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.688037 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.688057 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.688107 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.688123 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.688149 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790071 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790154 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w86zh\" (UniqueName: \"kubernetes.io/projected/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-kube-api-access-w86zh\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790196 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-b4484dbd-034d-48f1-a061-680e7edce5be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4484dbd-034d-48f1-a061-680e7edce5be\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790234 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790258 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790282 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790302 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.790337 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.791000 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.791345 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.800477 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-pod-info\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.800474 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.802491 4956 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.802536 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-b4484dbd-034d-48f1-a061-680e7edce5be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4484dbd-034d-48f1-a061-680e7edce5be\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/012beb37f3f7fce777c26cc21dc5196ece8b5b4d26f041ed4433c985a4c867e1/globalmount\"" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.806297 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.806921 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.826216 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w86zh\" (UniqueName: \"kubernetes.io/projected/d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312-kube-api-access-w86zh\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.832924 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-b4484dbd-034d-48f1-a061-680e7edce5be\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-b4484dbd-034d-48f1-a061-680e7edce5be\") pod \"rabbitmq-server-0\" (UID: \"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312\") " pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:06 crc kubenswrapper[4956]: I1126 17:08:06.896148 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:07 crc kubenswrapper[4956]: I1126 17:08:07.138176 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/rabbitmq-server-0"] Nov 26 17:08:07 crc kubenswrapper[4956]: I1126 17:08:07.758361 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312","Type":"ContainerStarted","Data":"1ab21215f55c0b576620fbcf5d4823168989a81367dd39b461085523985a1222"} Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.266432 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-index-phhw9"] Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.267418 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.271739 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-p4h5f" Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.277884 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-phhw9"] Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.415158 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zj7n9\" (UniqueName: \"kubernetes.io/projected/e70664dd-b470-4b7f-877c-8e7429a1ab7e-kube-api-access-zj7n9\") pod \"keystone-operator-index-phhw9\" (UID: \"e70664dd-b470-4b7f-877c-8e7429a1ab7e\") " pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.517368 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zj7n9\" (UniqueName: \"kubernetes.io/projected/e70664dd-b470-4b7f-877c-8e7429a1ab7e-kube-api-access-zj7n9\") pod \"keystone-operator-index-phhw9\" (UID: \"e70664dd-b470-4b7f-877c-8e7429a1ab7e\") " pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.539525 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zj7n9\" (UniqueName: \"kubernetes.io/projected/e70664dd-b470-4b7f-877c-8e7429a1ab7e-kube-api-access-zj7n9\") pod \"keystone-operator-index-phhw9\" (UID: \"e70664dd-b470-4b7f-877c-8e7429a1ab7e\") " pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:08 crc kubenswrapper[4956]: I1126 17:08:08.599998 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:09 crc kubenswrapper[4956]: I1126 17:08:09.021752 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-index-phhw9"] Nov 26 17:08:09 crc kubenswrapper[4956]: I1126 17:08:09.551680 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:08:09 crc kubenswrapper[4956]: I1126 17:08:09.551822 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:08:10 crc kubenswrapper[4956]: I1126 17:08:10.793853 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-phhw9" event={"ID":"e70664dd-b470-4b7f-877c-8e7429a1ab7e","Type":"ContainerStarted","Data":"09c070d6d8b8a9c34a8c54070dc4f455bd52b954fe4dee6b5535542d6490c522"} Nov 26 17:08:13 crc kubenswrapper[4956]: I1126 17:08:13.822596 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-index-phhw9" event={"ID":"e70664dd-b470-4b7f-877c-8e7429a1ab7e","Type":"ContainerStarted","Data":"26d73a1f5d1836e3c241f965ef0739d58ccac3866ee6e6f92c6120d0e8678eb3"} Nov 26 17:08:13 crc kubenswrapper[4956]: I1126 17:08:13.853259 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-index-phhw9" podStartSLOduration=2.330152395 podStartE2EDuration="5.85321837s" podCreationTimestamp="2025-11-26 17:08:08 +0000 UTC" firstStartedPulling="2025-11-26 17:08:09.937457165 +0000 UTC m=+935.633417717" lastFinishedPulling="2025-11-26 17:08:13.46052314 +0000 UTC m=+939.156483692" observedRunningTime="2025-11-26 17:08:13.842781242 +0000 UTC m=+939.538741794" watchObservedRunningTime="2025-11-26 17:08:13.85321837 +0000 UTC m=+939.549178962" Nov 26 17:08:14 crc kubenswrapper[4956]: I1126 17:08:14.835984 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312","Type":"ContainerStarted","Data":"9289142b3829db40eb891fe5eb039b204600c425aee344219f4dba0b837a0fe4"} Nov 26 17:08:18 crc kubenswrapper[4956]: I1126 17:08:18.600822 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:18 crc kubenswrapper[4956]: I1126 17:08:18.601275 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:18 crc kubenswrapper[4956]: I1126 17:08:18.635581 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:18 crc kubenswrapper[4956]: I1126 17:08:18.912569 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-index-phhw9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.710425 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9"] Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.712729 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.715921 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j97nd" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.723254 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9"] Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.863613 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.863736 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.863782 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95r4v\" (UniqueName: \"kubernetes.io/projected/1c6c002c-d6a2-444f-bd42-c2ba47045304-kube-api-access-95r4v\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.964828 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.964930 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95r4v\" (UniqueName: \"kubernetes.io/projected/1c6c002c-d6a2-444f-bd42-c2ba47045304-kube-api-access-95r4v\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.964996 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.965404 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-util\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.965498 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-bundle\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:27 crc kubenswrapper[4956]: I1126 17:08:27.999738 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95r4v\" (UniqueName: \"kubernetes.io/projected/1c6c002c-d6a2-444f-bd42-c2ba47045304-kube-api-access-95r4v\") pod \"d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:28 crc kubenswrapper[4956]: I1126 17:08:28.075281 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:28 crc kubenswrapper[4956]: I1126 17:08:28.332207 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9"] Nov 26 17:08:28 crc kubenswrapper[4956]: I1126 17:08:28.944637 4956 generic.go:334] "Generic (PLEG): container finished" podID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerID="ff43ff9f5a9ff7366a3e8450d3d0fbaef58c56efd203182ff989339b36d04373" exitCode=0 Nov 26 17:08:28 crc kubenswrapper[4956]: I1126 17:08:28.944718 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" event={"ID":"1c6c002c-d6a2-444f-bd42-c2ba47045304","Type":"ContainerDied","Data":"ff43ff9f5a9ff7366a3e8450d3d0fbaef58c56efd203182ff989339b36d04373"} Nov 26 17:08:28 crc kubenswrapper[4956]: I1126 17:08:28.944769 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" event={"ID":"1c6c002c-d6a2-444f-bd42-c2ba47045304","Type":"ContainerStarted","Data":"6b804fac6de0ab9cd7cffb5349cc6ade705d5e393da446de2b82fd79634e4c07"} Nov 26 17:08:30 crc kubenswrapper[4956]: I1126 17:08:30.963514 4956 generic.go:334] "Generic (PLEG): container finished" podID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerID="873da884e3f99280597012d8ca36813c359cabd08b3f2b4ec30cc5313d3ec47a" exitCode=0 Nov 26 17:08:30 crc kubenswrapper[4956]: I1126 17:08:30.963586 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" event={"ID":"1c6c002c-d6a2-444f-bd42-c2ba47045304","Type":"ContainerDied","Data":"873da884e3f99280597012d8ca36813c359cabd08b3f2b4ec30cc5313d3ec47a"} Nov 26 17:08:31 crc kubenswrapper[4956]: I1126 17:08:31.976491 4956 generic.go:334] "Generic (PLEG): container finished" podID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerID="35ad43cf3bdc045d4e12c9eb8db69afa2d3efb0191d7c5a5114219c8ed91973e" exitCode=0 Nov 26 17:08:31 crc kubenswrapper[4956]: I1126 17:08:31.976558 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" event={"ID":"1c6c002c-d6a2-444f-bd42-c2ba47045304","Type":"ContainerDied","Data":"35ad43cf3bdc045d4e12c9eb8db69afa2d3efb0191d7c5a5114219c8ed91973e"} Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.288729 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.447034 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-95r4v\" (UniqueName: \"kubernetes.io/projected/1c6c002c-d6a2-444f-bd42-c2ba47045304-kube-api-access-95r4v\") pod \"1c6c002c-d6a2-444f-bd42-c2ba47045304\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.447160 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-util\") pod \"1c6c002c-d6a2-444f-bd42-c2ba47045304\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.447281 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-bundle\") pod \"1c6c002c-d6a2-444f-bd42-c2ba47045304\" (UID: \"1c6c002c-d6a2-444f-bd42-c2ba47045304\") " Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.451229 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-bundle" (OuterVolumeSpecName: "bundle") pod "1c6c002c-d6a2-444f-bd42-c2ba47045304" (UID: "1c6c002c-d6a2-444f-bd42-c2ba47045304"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.456815 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c6c002c-d6a2-444f-bd42-c2ba47045304-kube-api-access-95r4v" (OuterVolumeSpecName: "kube-api-access-95r4v") pod "1c6c002c-d6a2-444f-bd42-c2ba47045304" (UID: "1c6c002c-d6a2-444f-bd42-c2ba47045304"). InnerVolumeSpecName "kube-api-access-95r4v". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.471414 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-util" (OuterVolumeSpecName: "util") pod "1c6c002c-d6a2-444f-bd42-c2ba47045304" (UID: "1c6c002c-d6a2-444f-bd42-c2ba47045304"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.549452 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-95r4v\" (UniqueName: \"kubernetes.io/projected/1c6c002c-d6a2-444f-bd42-c2ba47045304-kube-api-access-95r4v\") on node \"crc\" DevicePath \"\"" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.549807 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.549918 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/1c6c002c-d6a2-444f-bd42-c2ba47045304-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.996097 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" event={"ID":"1c6c002c-d6a2-444f-bd42-c2ba47045304","Type":"ContainerDied","Data":"6b804fac6de0ab9cd7cffb5349cc6ade705d5e393da446de2b82fd79634e4c07"} Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.996126 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9" Nov 26 17:08:33 crc kubenswrapper[4956]: I1126 17:08:33.996149 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6b804fac6de0ab9cd7cffb5349cc6ade705d5e393da446de2b82fd79634e4c07" Nov 26 17:08:39 crc kubenswrapper[4956]: I1126 17:08:39.551804 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:08:39 crc kubenswrapper[4956]: I1126 17:08:39.552481 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.826038 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf"] Nov 26 17:08:41 crc kubenswrapper[4956]: E1126 17:08:41.826774 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="pull" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.826792 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="pull" Nov 26 17:08:41 crc kubenswrapper[4956]: E1126 17:08:41.826806 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="extract" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.826812 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="extract" Nov 26 17:08:41 crc kubenswrapper[4956]: E1126 17:08:41.826832 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="util" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.826839 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="util" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.826997 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c6c002c-d6a2-444f-bd42-c2ba47045304" containerName="extract" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.827478 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.829514 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6jzcl" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.830748 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.849786 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf"] Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.895974 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f723082-bb1f-4dd1-beb2-c1553728eba8-webhook-cert\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.896041 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/7f723082-bb1f-4dd1-beb2-c1553728eba8-kube-api-access-h8bzv\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.896068 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f723082-bb1f-4dd1-beb2-c1553728eba8-apiservice-cert\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.997794 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f723082-bb1f-4dd1-beb2-c1553728eba8-webhook-cert\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.998157 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/7f723082-bb1f-4dd1-beb2-c1553728eba8-kube-api-access-h8bzv\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:41 crc kubenswrapper[4956]: I1126 17:08:41.998249 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f723082-bb1f-4dd1-beb2-c1553728eba8-apiservice-cert\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:42 crc kubenswrapper[4956]: I1126 17:08:42.020935 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/7f723082-bb1f-4dd1-beb2-c1553728eba8-webhook-cert\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:42 crc kubenswrapper[4956]: I1126 17:08:42.029974 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h8bzv\" (UniqueName: \"kubernetes.io/projected/7f723082-bb1f-4dd1-beb2-c1553728eba8-kube-api-access-h8bzv\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:42 crc kubenswrapper[4956]: I1126 17:08:42.030669 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/7f723082-bb1f-4dd1-beb2-c1553728eba8-apiservice-cert\") pod \"keystone-operator-controller-manager-69b487756b-xzvnf\" (UID: \"7f723082-bb1f-4dd1-beb2-c1553728eba8\") " pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:42 crc kubenswrapper[4956]: I1126 17:08:42.146048 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:42 crc kubenswrapper[4956]: I1126 17:08:42.402369 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf"] Nov 26 17:08:42 crc kubenswrapper[4956]: I1126 17:08:42.420310 4956 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 17:08:43 crc kubenswrapper[4956]: I1126 17:08:43.068423 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" event={"ID":"7f723082-bb1f-4dd1-beb2-c1553728eba8","Type":"ContainerStarted","Data":"1237ba06ecaf83beb74a1d9011096908e070fd12ad2ef16669c28697b4a3f68a"} Nov 26 17:08:47 crc kubenswrapper[4956]: I1126 17:08:47.110213 4956 generic.go:334] "Generic (PLEG): container finished" podID="d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312" containerID="9289142b3829db40eb891fe5eb039b204600c425aee344219f4dba0b837a0fe4" exitCode=0 Nov 26 17:08:47 crc kubenswrapper[4956]: I1126 17:08:47.110290 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312","Type":"ContainerDied","Data":"9289142b3829db40eb891fe5eb039b204600c425aee344219f4dba0b837a0fe4"} Nov 26 17:08:47 crc kubenswrapper[4956]: I1126 17:08:47.114724 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" event={"ID":"7f723082-bb1f-4dd1-beb2-c1553728eba8","Type":"ContainerStarted","Data":"3a76257a664da6b026934dc765eb9fb3c30538c411787a826217022c03664890"} Nov 26 17:08:47 crc kubenswrapper[4956]: I1126 17:08:47.114960 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:48 crc kubenswrapper[4956]: I1126 17:08:48.123757 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/rabbitmq-server-0" event={"ID":"d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312","Type":"ContainerStarted","Data":"5577a44bd3ebe187fa08cb0adaf686dcd3e41767517afc83fe76fce2579645c3"} Nov 26 17:08:48 crc kubenswrapper[4956]: I1126 17:08:48.124617 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:08:48 crc kubenswrapper[4956]: I1126 17:08:48.149072 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" podStartSLOduration=3.050156535 podStartE2EDuration="7.149019012s" podCreationTimestamp="2025-11-26 17:08:41 +0000 UTC" firstStartedPulling="2025-11-26 17:08:42.420058104 +0000 UTC m=+968.116018646" lastFinishedPulling="2025-11-26 17:08:46.518920571 +0000 UTC m=+972.214881123" observedRunningTime="2025-11-26 17:08:47.213983864 +0000 UTC m=+972.909944426" watchObservedRunningTime="2025-11-26 17:08:48.149019012 +0000 UTC m=+973.844979564" Nov 26 17:08:48 crc kubenswrapper[4956]: I1126 17:08:48.152500 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/rabbitmq-server-0" podStartSLOduration=37.115327805 podStartE2EDuration="43.152487521s" podCreationTimestamp="2025-11-26 17:08:05 +0000 UTC" firstStartedPulling="2025-11-26 17:08:07.1395014 +0000 UTC m=+932.835461962" lastFinishedPulling="2025-11-26 17:08:13.176661126 +0000 UTC m=+938.872621678" observedRunningTime="2025-11-26 17:08:48.145883412 +0000 UTC m=+973.841843984" watchObservedRunningTime="2025-11-26 17:08:48.152487521 +0000 UTC m=+973.848448073" Nov 26 17:08:52 crc kubenswrapper[4956]: I1126 17:08:52.153080 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.857741 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb"] Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.859093 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.861782 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-db-secret" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.865377 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-create-76k8h"] Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.866338 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.884314 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb"] Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.900840 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-76k8h"] Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.927717 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a79036d2-2c7e-4b04-a1bc-b194074b27e7-operator-scripts\") pod \"keystone-db-create-76k8h\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.927792 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w4ld7\" (UniqueName: \"kubernetes.io/projected/a79036d2-2c7e-4b04-a1bc-b194074b27e7-kube-api-access-w4ld7\") pod \"keystone-db-create-76k8h\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.927835 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be4a1c0e-bb65-4419-9349-708ffe694948-operator-scripts\") pod \"keystone-8bae-account-create-update-8x9jb\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:55 crc kubenswrapper[4956]: I1126 17:08:55.927901 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tvfkp\" (UniqueName: \"kubernetes.io/projected/be4a1c0e-bb65-4419-9349-708ffe694948-kube-api-access-tvfkp\") pod \"keystone-8bae-account-create-update-8x9jb\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.029094 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a79036d2-2c7e-4b04-a1bc-b194074b27e7-operator-scripts\") pod \"keystone-db-create-76k8h\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.029160 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w4ld7\" (UniqueName: \"kubernetes.io/projected/a79036d2-2c7e-4b04-a1bc-b194074b27e7-kube-api-access-w4ld7\") pod \"keystone-db-create-76k8h\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.029210 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be4a1c0e-bb65-4419-9349-708ffe694948-operator-scripts\") pod \"keystone-8bae-account-create-update-8x9jb\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.029257 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tvfkp\" (UniqueName: \"kubernetes.io/projected/be4a1c0e-bb65-4419-9349-708ffe694948-kube-api-access-tvfkp\") pod \"keystone-8bae-account-create-update-8x9jb\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.030082 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a79036d2-2c7e-4b04-a1bc-b194074b27e7-operator-scripts\") pod \"keystone-db-create-76k8h\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.030098 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be4a1c0e-bb65-4419-9349-708ffe694948-operator-scripts\") pod \"keystone-8bae-account-create-update-8x9jb\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.050782 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w4ld7\" (UniqueName: \"kubernetes.io/projected/a79036d2-2c7e-4b04-a1bc-b194074b27e7-kube-api-access-w4ld7\") pod \"keystone-db-create-76k8h\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.057457 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tvfkp\" (UniqueName: \"kubernetes.io/projected/be4a1c0e-bb65-4419-9349-708ffe694948-kube-api-access-tvfkp\") pod \"keystone-8bae-account-create-update-8x9jb\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.188607 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.195225 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.740590 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb"] Nov 26 17:08:56 crc kubenswrapper[4956]: I1126 17:08:56.758468 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-create-76k8h"] Nov 26 17:08:56 crc kubenswrapper[4956]: W1126 17:08:56.773822 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda79036d2_2c7e_4b04_a1bc_b194074b27e7.slice/crio-049c36d479c74533de16ceb0442cbd6415737e349f30363f70c3c307d6af5b50 WatchSource:0}: Error finding container 049c36d479c74533de16ceb0442cbd6415737e349f30363f70c3c307d6af5b50: Status 404 returned error can't find the container with id 049c36d479c74533de16ceb0442cbd6415737e349f30363f70c3c307d6af5b50 Nov 26 17:08:57 crc kubenswrapper[4956]: I1126 17:08:57.188191 4956 generic.go:334] "Generic (PLEG): container finished" podID="a79036d2-2c7e-4b04-a1bc-b194074b27e7" containerID="62e61e25806244bd7858bc61f6f04696cce99aa4c947ba0fda6486fc883f4ed0" exitCode=0 Nov 26 17:08:57 crc kubenswrapper[4956]: I1126 17:08:57.188308 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-76k8h" event={"ID":"a79036d2-2c7e-4b04-a1bc-b194074b27e7","Type":"ContainerDied","Data":"62e61e25806244bd7858bc61f6f04696cce99aa4c947ba0fda6486fc883f4ed0"} Nov 26 17:08:57 crc kubenswrapper[4956]: I1126 17:08:57.188681 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-76k8h" event={"ID":"a79036d2-2c7e-4b04-a1bc-b194074b27e7","Type":"ContainerStarted","Data":"049c36d479c74533de16ceb0442cbd6415737e349f30363f70c3c307d6af5b50"} Nov 26 17:08:57 crc kubenswrapper[4956]: I1126 17:08:57.190736 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" event={"ID":"be4a1c0e-bb65-4419-9349-708ffe694948","Type":"ContainerStarted","Data":"0810e3706926c9d3f1498390eb1e795e8d17c4af87fa04dbdc4932d5ff1ca150"} Nov 26 17:08:57 crc kubenswrapper[4956]: I1126 17:08:57.190779 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" event={"ID":"be4a1c0e-bb65-4419-9349-708ffe694948","Type":"ContainerStarted","Data":"0a331bd7c0d18b5e4b6ede8efbabfe8c2869a7fe129deecd8cf11d67375e072a"} Nov 26 17:08:57 crc kubenswrapper[4956]: I1126 17:08:57.228730 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" podStartSLOduration=2.228703324 podStartE2EDuration="2.228703324s" podCreationTimestamp="2025-11-26 17:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:08:57.225546684 +0000 UTC m=+982.921507256" watchObservedRunningTime="2025-11-26 17:08:57.228703324 +0000 UTC m=+982.924663876" Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.200410 4956 generic.go:334] "Generic (PLEG): container finished" podID="be4a1c0e-bb65-4419-9349-708ffe694948" containerID="0810e3706926c9d3f1498390eb1e795e8d17c4af87fa04dbdc4932d5ff1ca150" exitCode=0 Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.200539 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" event={"ID":"be4a1c0e-bb65-4419-9349-708ffe694948","Type":"ContainerDied","Data":"0810e3706926c9d3f1498390eb1e795e8d17c4af87fa04dbdc4932d5ff1ca150"} Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.485567 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.610038 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a79036d2-2c7e-4b04-a1bc-b194074b27e7-operator-scripts\") pod \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.610893 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a79036d2-2c7e-4b04-a1bc-b194074b27e7-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "a79036d2-2c7e-4b04-a1bc-b194074b27e7" (UID: "a79036d2-2c7e-4b04-a1bc-b194074b27e7"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.610895 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4ld7\" (UniqueName: \"kubernetes.io/projected/a79036d2-2c7e-4b04-a1bc-b194074b27e7-kube-api-access-w4ld7\") pod \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\" (UID: \"a79036d2-2c7e-4b04-a1bc-b194074b27e7\") " Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.611572 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/a79036d2-2c7e-4b04-a1bc-b194074b27e7-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.622306 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a79036d2-2c7e-4b04-a1bc-b194074b27e7-kube-api-access-w4ld7" (OuterVolumeSpecName: "kube-api-access-w4ld7") pod "a79036d2-2c7e-4b04-a1bc-b194074b27e7" (UID: "a79036d2-2c7e-4b04-a1bc-b194074b27e7"). InnerVolumeSpecName "kube-api-access-w4ld7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:08:58 crc kubenswrapper[4956]: I1126 17:08:58.712819 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4ld7\" (UniqueName: \"kubernetes.io/projected/a79036d2-2c7e-4b04-a1bc-b194074b27e7-kube-api-access-w4ld7\") on node \"crc\" DevicePath \"\"" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.212353 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-create-76k8h" event={"ID":"a79036d2-2c7e-4b04-a1bc-b194074b27e7","Type":"ContainerDied","Data":"049c36d479c74533de16ceb0442cbd6415737e349f30363f70c3c307d6af5b50"} Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.212429 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="049c36d479c74533de16ceb0442cbd6415737e349f30363f70c3c307d6af5b50" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.212458 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-create-76k8h" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.510451 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.626931 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tvfkp\" (UniqueName: \"kubernetes.io/projected/be4a1c0e-bb65-4419-9349-708ffe694948-kube-api-access-tvfkp\") pod \"be4a1c0e-bb65-4419-9349-708ffe694948\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.627018 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be4a1c0e-bb65-4419-9349-708ffe694948-operator-scripts\") pod \"be4a1c0e-bb65-4419-9349-708ffe694948\" (UID: \"be4a1c0e-bb65-4419-9349-708ffe694948\") " Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.628122 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/be4a1c0e-bb65-4419-9349-708ffe694948-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "be4a1c0e-bb65-4419-9349-708ffe694948" (UID: "be4a1c0e-bb65-4419-9349-708ffe694948"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.635571 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be4a1c0e-bb65-4419-9349-708ffe694948-kube-api-access-tvfkp" (OuterVolumeSpecName: "kube-api-access-tvfkp") pod "be4a1c0e-bb65-4419-9349-708ffe694948" (UID: "be4a1c0e-bb65-4419-9349-708ffe694948"). InnerVolumeSpecName "kube-api-access-tvfkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.729305 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tvfkp\" (UniqueName: \"kubernetes.io/projected/be4a1c0e-bb65-4419-9349-708ffe694948-kube-api-access-tvfkp\") on node \"crc\" DevicePath \"\"" Nov 26 17:08:59 crc kubenswrapper[4956]: I1126 17:08:59.729343 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/be4a1c0e-bb65-4419-9349-708ffe694948-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:00 crc kubenswrapper[4956]: I1126 17:09:00.220422 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" event={"ID":"be4a1c0e-bb65-4419-9349-708ffe694948","Type":"ContainerDied","Data":"0a331bd7c0d18b5e4b6ede8efbabfe8c2869a7fe129deecd8cf11d67375e072a"} Nov 26 17:09:00 crc kubenswrapper[4956]: I1126 17:09:00.220473 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0a331bd7c0d18b5e4b6ede8efbabfe8c2869a7fe129deecd8cf11d67375e072a" Nov 26 17:09:00 crc kubenswrapper[4956]: I1126 17:09:00.220548 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb" Nov 26 17:09:06 crc kubenswrapper[4956]: I1126 17:09:06.901561 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/rabbitmq-server-0" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.459107 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7866q"] Nov 26 17:09:07 crc kubenswrapper[4956]: E1126 17:09:07.459420 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be4a1c0e-bb65-4419-9349-708ffe694948" containerName="mariadb-account-create-update" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.459433 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="be4a1c0e-bb65-4419-9349-708ffe694948" containerName="mariadb-account-create-update" Nov 26 17:09:07 crc kubenswrapper[4956]: E1126 17:09:07.459450 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a79036d2-2c7e-4b04-a1bc-b194074b27e7" containerName="mariadb-database-create" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.459457 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a79036d2-2c7e-4b04-a1bc-b194074b27e7" containerName="mariadb-database-create" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.459574 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="be4a1c0e-bb65-4419-9349-708ffe694948" containerName="mariadb-account-create-update" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.459586 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="a79036d2-2c7e-4b04-a1bc-b194074b27e7" containerName="mariadb-database-create" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.460132 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.462759 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.463335 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.463465 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.463664 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-w6kjs" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.472790 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7866q"] Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.558050 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xhnh2\" (UniqueName: \"kubernetes.io/projected/8d16e0ee-c3be-4697-a639-a3274721d828-kube-api-access-xhnh2\") pod \"keystone-db-sync-7866q\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.558445 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d16e0ee-c3be-4697-a639-a3274721d828-config-data\") pod \"keystone-db-sync-7866q\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.660560 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xhnh2\" (UniqueName: \"kubernetes.io/projected/8d16e0ee-c3be-4697-a639-a3274721d828-kube-api-access-xhnh2\") pod \"keystone-db-sync-7866q\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.660643 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d16e0ee-c3be-4697-a639-a3274721d828-config-data\") pod \"keystone-db-sync-7866q\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.669728 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d16e0ee-c3be-4697-a639-a3274721d828-config-data\") pod \"keystone-db-sync-7866q\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.679209 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xhnh2\" (UniqueName: \"kubernetes.io/projected/8d16e0ee-c3be-4697-a639-a3274721d828-kube-api-access-xhnh2\") pod \"keystone-db-sync-7866q\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:07 crc kubenswrapper[4956]: I1126 17:09:07.802315 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.058120 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7866q"] Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.287050 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7866q" event={"ID":"8d16e0ee-c3be-4697-a639-a3274721d828","Type":"ContainerStarted","Data":"c31aaa67c5ab6341b2e533bc1bf5f18e243fe2a017293f01cfa0dffb84564950"} Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.873842 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-index-kb7vl"] Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.876365 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.881232 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-xfgqg" Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.885620 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-kb7vl"] Nov 26 17:09:08 crc kubenswrapper[4956]: I1126 17:09:08.985953 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgtmh\" (UniqueName: \"kubernetes.io/projected/f9b0e47e-e22e-4b2e-8b8c-732bb510c62d-kube-api-access-tgtmh\") pod \"horizon-operator-index-kb7vl\" (UID: \"f9b0e47e-e22e-4b2e-8b8c-732bb510c62d\") " pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.088104 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgtmh\" (UniqueName: \"kubernetes.io/projected/f9b0e47e-e22e-4b2e-8b8c-732bb510c62d-kube-api-access-tgtmh\") pod \"horizon-operator-index-kb7vl\" (UID: \"f9b0e47e-e22e-4b2e-8b8c-732bb510c62d\") " pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.112205 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgtmh\" (UniqueName: \"kubernetes.io/projected/f9b0e47e-e22e-4b2e-8b8c-732bb510c62d-kube-api-access-tgtmh\") pod \"horizon-operator-index-kb7vl\" (UID: \"f9b0e47e-e22e-4b2e-8b8c-732bb510c62d\") " pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.212096 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.444157 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-index-kb7vl"] Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.551006 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.551099 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.551163 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.551932 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"0683eafa4a97fe7cebdb5b35d5b4a8e27469c373b19946cd7cb9d38e21f774da"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 17:09:09 crc kubenswrapper[4956]: I1126 17:09:09.551993 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://0683eafa4a97fe7cebdb5b35d5b4a8e27469c373b19946cd7cb9d38e21f774da" gracePeriod=600 Nov 26 17:09:10 crc kubenswrapper[4956]: I1126 17:09:10.308123 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-kb7vl" event={"ID":"f9b0e47e-e22e-4b2e-8b8c-732bb510c62d","Type":"ContainerStarted","Data":"161c29223493b9589558d65b51426caa8b652e201237e53954bed605ab9d85fc"} Nov 26 17:09:10 crc kubenswrapper[4956]: I1126 17:09:10.312988 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="0683eafa4a97fe7cebdb5b35d5b4a8e27469c373b19946cd7cb9d38e21f774da" exitCode=0 Nov 26 17:09:10 crc kubenswrapper[4956]: I1126 17:09:10.313076 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"0683eafa4a97fe7cebdb5b35d5b4a8e27469c373b19946cd7cb9d38e21f774da"} Nov 26 17:09:10 crc kubenswrapper[4956]: I1126 17:09:10.313167 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"e26914562ac9efcf004522ee5db12eb4d489e28d6cbf474316c25c32bd557c51"} Nov 26 17:09:10 crc kubenswrapper[4956]: I1126 17:09:10.313189 4956 scope.go:117] "RemoveContainer" containerID="60fb36ae8e9dbdf5a9ddd39194e72f5c72fcc2005c642dd846c7d001b368d385" Nov 26 17:09:11 crc kubenswrapper[4956]: I1126 17:09:11.326216 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-index-kb7vl" event={"ID":"f9b0e47e-e22e-4b2e-8b8c-732bb510c62d","Type":"ContainerStarted","Data":"527988059a8f4fd665e2e6b603cb5614c25fda1854d937c61f54d132e127f3fb"} Nov 26 17:09:13 crc kubenswrapper[4956]: I1126 17:09:13.870602 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-index-kb7vl" podStartSLOduration=4.936513245 podStartE2EDuration="5.870579672s" podCreationTimestamp="2025-11-26 17:09:08 +0000 UTC" firstStartedPulling="2025-11-26 17:09:09.467040328 +0000 UTC m=+995.163000880" lastFinishedPulling="2025-11-26 17:09:10.401106755 +0000 UTC m=+996.097067307" observedRunningTime="2025-11-26 17:09:11.347596415 +0000 UTC m=+997.043556977" watchObservedRunningTime="2025-11-26 17:09:13.870579672 +0000 UTC m=+999.566540234" Nov 26 17:09:13 crc kubenswrapper[4956]: I1126 17:09:13.873839 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-index-rwrnv"] Nov 26 17:09:13 crc kubenswrapper[4956]: I1126 17:09:13.874804 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:13 crc kubenswrapper[4956]: I1126 17:09:13.879415 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-s25tb" Nov 26 17:09:13 crc kubenswrapper[4956]: I1126 17:09:13.884529 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-rwrnv"] Nov 26 17:09:13 crc kubenswrapper[4956]: I1126 17:09:13.973764 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c2bsf\" (UniqueName: \"kubernetes.io/projected/434482ad-0b5b-47a5-aa36-289139e64252-kube-api-access-c2bsf\") pod \"swift-operator-index-rwrnv\" (UID: \"434482ad-0b5b-47a5-aa36-289139e64252\") " pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:14 crc kubenswrapper[4956]: I1126 17:09:14.075815 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c2bsf\" (UniqueName: \"kubernetes.io/projected/434482ad-0b5b-47a5-aa36-289139e64252-kube-api-access-c2bsf\") pod \"swift-operator-index-rwrnv\" (UID: \"434482ad-0b5b-47a5-aa36-289139e64252\") " pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:14 crc kubenswrapper[4956]: I1126 17:09:14.108687 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c2bsf\" (UniqueName: \"kubernetes.io/projected/434482ad-0b5b-47a5-aa36-289139e64252-kube-api-access-c2bsf\") pod \"swift-operator-index-rwrnv\" (UID: \"434482ad-0b5b-47a5-aa36-289139e64252\") " pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:14 crc kubenswrapper[4956]: I1126 17:09:14.202973 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:19 crc kubenswrapper[4956]: I1126 17:09:19.213324 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:19 crc kubenswrapper[4956]: I1126 17:09:19.214176 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:19 crc kubenswrapper[4956]: I1126 17:09:19.247379 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:19 crc kubenswrapper[4956]: I1126 17:09:19.423538 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-index-kb7vl" Nov 26 17:09:20 crc kubenswrapper[4956]: I1126 17:09:20.397811 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7866q" event={"ID":"8d16e0ee-c3be-4697-a639-a3274721d828","Type":"ContainerStarted","Data":"648e44c093b0e44c9efb0fcc198e73caa1ef6477ce350814cc0a7a8dc82621c0"} Nov 26 17:09:20 crc kubenswrapper[4956]: I1126 17:09:20.430612 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-db-sync-7866q" podStartSLOduration=1.461048177 podStartE2EDuration="13.430590844s" podCreationTimestamp="2025-11-26 17:09:07 +0000 UTC" firstStartedPulling="2025-11-26 17:09:08.067285841 +0000 UTC m=+993.763246383" lastFinishedPulling="2025-11-26 17:09:20.036828498 +0000 UTC m=+1005.732789050" observedRunningTime="2025-11-26 17:09:20.425076957 +0000 UTC m=+1006.121037519" watchObservedRunningTime="2025-11-26 17:09:20.430590844 +0000 UTC m=+1006.126551396" Nov 26 17:09:20 crc kubenswrapper[4956]: I1126 17:09:20.483924 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-index-rwrnv"] Nov 26 17:09:20 crc kubenswrapper[4956]: W1126 17:09:20.495392 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod434482ad_0b5b_47a5_aa36_289139e64252.slice/crio-0f96ae452dbe82b15e982b6dbb4f63ebc51d9696baca0a083490a81f138a2ad6 WatchSource:0}: Error finding container 0f96ae452dbe82b15e982b6dbb4f63ebc51d9696baca0a083490a81f138a2ad6: Status 404 returned error can't find the container with id 0f96ae452dbe82b15e982b6dbb4f63ebc51d9696baca0a083490a81f138a2ad6 Nov 26 17:09:21 crc kubenswrapper[4956]: I1126 17:09:21.408467 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-rwrnv" event={"ID":"434482ad-0b5b-47a5-aa36-289139e64252","Type":"ContainerStarted","Data":"0f96ae452dbe82b15e982b6dbb4f63ebc51d9696baca0a083490a81f138a2ad6"} Nov 26 17:09:22 crc kubenswrapper[4956]: I1126 17:09:22.416437 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-index-rwrnv" event={"ID":"434482ad-0b5b-47a5-aa36-289139e64252","Type":"ContainerStarted","Data":"2513acf73861411927d91cdfab11a35d115ab87ec1962548906f994f8a15779f"} Nov 26 17:09:22 crc kubenswrapper[4956]: I1126 17:09:22.436732 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-index-rwrnv" podStartSLOduration=8.415953236 podStartE2EDuration="9.436712088s" podCreationTimestamp="2025-11-26 17:09:13 +0000 UTC" firstStartedPulling="2025-11-26 17:09:20.497534816 +0000 UTC m=+1006.193495368" lastFinishedPulling="2025-11-26 17:09:21.518293668 +0000 UTC m=+1007.214254220" observedRunningTime="2025-11-26 17:09:22.430768938 +0000 UTC m=+1008.126729510" watchObservedRunningTime="2025-11-26 17:09:22.436712088 +0000 UTC m=+1008.132672640" Nov 26 17:09:24 crc kubenswrapper[4956]: I1126 17:09:24.203431 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:24 crc kubenswrapper[4956]: I1126 17:09:24.204103 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:24 crc kubenswrapper[4956]: I1126 17:09:24.244766 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:24 crc kubenswrapper[4956]: I1126 17:09:24.436404 4956 generic.go:334] "Generic (PLEG): container finished" podID="8d16e0ee-c3be-4697-a639-a3274721d828" containerID="648e44c093b0e44c9efb0fcc198e73caa1ef6477ce350814cc0a7a8dc82621c0" exitCode=0 Nov 26 17:09:24 crc kubenswrapper[4956]: I1126 17:09:24.436512 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7866q" event={"ID":"8d16e0ee-c3be-4697-a639-a3274721d828","Type":"ContainerDied","Data":"648e44c093b0e44c9efb0fcc198e73caa1ef6477ce350814cc0a7a8dc82621c0"} Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.719820 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.819299 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xhnh2\" (UniqueName: \"kubernetes.io/projected/8d16e0ee-c3be-4697-a639-a3274721d828-kube-api-access-xhnh2\") pod \"8d16e0ee-c3be-4697-a639-a3274721d828\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.819361 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d16e0ee-c3be-4697-a639-a3274721d828-config-data\") pod \"8d16e0ee-c3be-4697-a639-a3274721d828\" (UID: \"8d16e0ee-c3be-4697-a639-a3274721d828\") " Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.828099 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8d16e0ee-c3be-4697-a639-a3274721d828-kube-api-access-xhnh2" (OuterVolumeSpecName: "kube-api-access-xhnh2") pod "8d16e0ee-c3be-4697-a639-a3274721d828" (UID: "8d16e0ee-c3be-4697-a639-a3274721d828"). InnerVolumeSpecName "kube-api-access-xhnh2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.862801 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8d16e0ee-c3be-4697-a639-a3274721d828-config-data" (OuterVolumeSpecName: "config-data") pod "8d16e0ee-c3be-4697-a639-a3274721d828" (UID: "8d16e0ee-c3be-4697-a639-a3274721d828"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.920617 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xhnh2\" (UniqueName: \"kubernetes.io/projected/8d16e0ee-c3be-4697-a639-a3274721d828-kube-api-access-xhnh2\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:25 crc kubenswrapper[4956]: I1126 17:09:25.920661 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8d16e0ee-c3be-4697-a639-a3274721d828-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.454286 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-db-sync-7866q" event={"ID":"8d16e0ee-c3be-4697-a639-a3274721d828","Type":"ContainerDied","Data":"c31aaa67c5ab6341b2e533bc1bf5f18e243fe2a017293f01cfa0dffb84564950"} Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.454342 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c31aaa67c5ab6341b2e533bc1bf5f18e243fe2a017293f01cfa0dffb84564950" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.454348 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-db-sync-7866q" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.705770 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6txpz"] Nov 26 17:09:26 crc kubenswrapper[4956]: E1126 17:09:26.706110 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8d16e0ee-c3be-4697-a639-a3274721d828" containerName="keystone-db-sync" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.706126 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8d16e0ee-c3be-4697-a639-a3274721d828" containerName="keystone-db-sync" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.706247 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8d16e0ee-c3be-4697-a639-a3274721d828" containerName="keystone-db-sync" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.706894 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.711537 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.711622 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.711681 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-w6kjs" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.711890 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.711895 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"osp-secret" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.762245 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6txpz"] Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.872981 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-fernet-keys\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.873059 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-scripts\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.873737 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sn54r\" (UniqueName: \"kubernetes.io/projected/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-kube-api-access-sn54r\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.873852 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-config-data\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.873986 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-credential-keys\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.975220 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sn54r\" (UniqueName: \"kubernetes.io/projected/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-kube-api-access-sn54r\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.975280 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-config-data\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.975321 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-credential-keys\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.975346 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-fernet-keys\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.975377 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-scripts\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.980457 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-scripts\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.980612 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-credential-keys\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.980778 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-fernet-keys\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:26 crc kubenswrapper[4956]: I1126 17:09:26.980933 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-config-data\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:27 crc kubenswrapper[4956]: I1126 17:09:27.009697 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sn54r\" (UniqueName: \"kubernetes.io/projected/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-kube-api-access-sn54r\") pod \"keystone-bootstrap-6txpz\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:27 crc kubenswrapper[4956]: I1126 17:09:27.024855 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:27 crc kubenswrapper[4956]: I1126 17:09:27.511415 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6txpz"] Nov 26 17:09:28 crc kubenswrapper[4956]: I1126 17:09:28.471298 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" event={"ID":"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0","Type":"ContainerStarted","Data":"d7219e7cc62d280e0b3569ebf752e6643236c7aed32e34571bee8d4f49a9ada9"} Nov 26 17:09:28 crc kubenswrapper[4956]: I1126 17:09:28.471692 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" event={"ID":"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0","Type":"ContainerStarted","Data":"e36f63cd3845f8df6e4306c357a8467a0892c36a1a592467de4830fff644ab66"} Nov 26 17:09:28 crc kubenswrapper[4956]: I1126 17:09:28.505136 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" podStartSLOduration=2.505106879 podStartE2EDuration="2.505106879s" podCreationTimestamp="2025-11-26 17:09:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:09:28.502584357 +0000 UTC m=+1014.198544949" watchObservedRunningTime="2025-11-26 17:09:28.505106879 +0000 UTC m=+1014.201067471" Nov 26 17:09:31 crc kubenswrapper[4956]: I1126 17:09:31.493133 4956 generic.go:334] "Generic (PLEG): container finished" podID="a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" containerID="d7219e7cc62d280e0b3569ebf752e6643236c7aed32e34571bee8d4f49a9ada9" exitCode=0 Nov 26 17:09:31 crc kubenswrapper[4956]: I1126 17:09:31.493212 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" event={"ID":"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0","Type":"ContainerDied","Data":"d7219e7cc62d280e0b3569ebf752e6643236c7aed32e34571bee8d4f49a9ada9"} Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.825767 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.962531 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sn54r\" (UniqueName: \"kubernetes.io/projected/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-kube-api-access-sn54r\") pod \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.962659 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-credential-keys\") pod \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.962716 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-scripts\") pod \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.962778 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-config-data\") pod \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.962818 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-fernet-keys\") pod \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\" (UID: \"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0\") " Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.969351 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-scripts" (OuterVolumeSpecName: "scripts") pod "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" (UID: "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.970045 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" (UID: "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.970424 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" (UID: "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.971201 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-kube-api-access-sn54r" (OuterVolumeSpecName: "kube-api-access-sn54r") pod "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" (UID: "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0"). InnerVolumeSpecName "kube-api-access-sn54r". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:09:32 crc kubenswrapper[4956]: I1126 17:09:32.985413 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-config-data" (OuterVolumeSpecName: "config-data") pod "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" (UID: "a096a446-ae0a-4e84-88a6-2fe7d8ca35f0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.064659 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sn54r\" (UniqueName: \"kubernetes.io/projected/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-kube-api-access-sn54r\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.065098 4956 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-credential-keys\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.065372 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.065529 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.065662 4956 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0-fernet-keys\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.511676 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" event={"ID":"a096a446-ae0a-4e84-88a6-2fe7d8ca35f0","Type":"ContainerDied","Data":"e36f63cd3845f8df6e4306c357a8467a0892c36a1a592467de4830fff644ab66"} Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.512121 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e36f63cd3845f8df6e4306c357a8467a0892c36a1a592467de4830fff644ab66" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.511725 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-bootstrap-6txpz" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.606318 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/keystone-6c49f4fcbc-tchbn"] Nov 26 17:09:33 crc kubenswrapper[4956]: E1126 17:09:33.606925 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" containerName="keystone-bootstrap" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.607005 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" containerName="keystone-bootstrap" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.607187 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" containerName="keystone-bootstrap" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.607735 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.610565 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.611756 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.611785 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-w6kjs" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.612093 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.624326 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-6c49f4fcbc-tchbn"] Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.776739 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sqmmk\" (UniqueName: \"kubernetes.io/projected/6520d055-1a7e-47f7-a3c6-08d59d991dd8-kube-api-access-sqmmk\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.776822 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-credential-keys\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.776901 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-fernet-keys\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.776964 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-config-data\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.777033 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-scripts\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.879553 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-credential-keys\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.879622 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-fernet-keys\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.879678 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-config-data\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.879727 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-scripts\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.885432 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-scripts\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.885614 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sqmmk\" (UniqueName: \"kubernetes.io/projected/6520d055-1a7e-47f7-a3c6-08d59d991dd8-kube-api-access-sqmmk\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.887635 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-fernet-keys\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.892215 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-credential-keys\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.902771 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6520d055-1a7e-47f7-a3c6-08d59d991dd8-config-data\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.909401 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sqmmk\" (UniqueName: \"kubernetes.io/projected/6520d055-1a7e-47f7-a3c6-08d59d991dd8-kube-api-access-sqmmk\") pod \"keystone-6c49f4fcbc-tchbn\" (UID: \"6520d055-1a7e-47f7-a3c6-08d59d991dd8\") " pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:33 crc kubenswrapper[4956]: I1126 17:09:33.923690 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:34 crc kubenswrapper[4956]: I1126 17:09:34.158301 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/keystone-6c49f4fcbc-tchbn"] Nov 26 17:09:34 crc kubenswrapper[4956]: I1126 17:09:34.238359 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-index-rwrnv" Nov 26 17:09:34 crc kubenswrapper[4956]: I1126 17:09:34.519431 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" event={"ID":"6520d055-1a7e-47f7-a3c6-08d59d991dd8","Type":"ContainerStarted","Data":"da9524b9483971ae6c8af4518d03e8cff3a506d307808cdcb26f4ea5e7eb40a6"} Nov 26 17:09:34 crc kubenswrapper[4956]: I1126 17:09:34.519481 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" event={"ID":"6520d055-1a7e-47f7-a3c6-08d59d991dd8","Type":"ContainerStarted","Data":"bdbed3451cd313e1f1cb21c7eace4465938300136455c053291dc472a2c52159"} Nov 26 17:09:34 crc kubenswrapper[4956]: I1126 17:09:34.519601 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:09:34 crc kubenswrapper[4956]: I1126 17:09:34.537613 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" podStartSLOduration=1.537595306 podStartE2EDuration="1.537595306s" podCreationTimestamp="2025-11-26 17:09:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:09:34.537102902 +0000 UTC m=+1020.233063454" watchObservedRunningTime="2025-11-26 17:09:34.537595306 +0000 UTC m=+1020.233555858" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.714685 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq"] Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.717984 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.720653 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j97nd" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.730196 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq"] Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.813793 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.814109 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-br4tl\" (UniqueName: \"kubernetes.io/projected/71a5370b-ba54-41ae-8e74-38a191f344d0-kube-api-access-br4tl\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.814159 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.915678 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.915787 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.915854 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-br4tl\" (UniqueName: \"kubernetes.io/projected/71a5370b-ba54-41ae-8e74-38a191f344d0-kube-api-access-br4tl\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.916437 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-bundle\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.916536 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-util\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:40 crc kubenswrapper[4956]: I1126 17:09:40.941970 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-br4tl\" (UniqueName: \"kubernetes.io/projected/71a5370b-ba54-41ae-8e74-38a191f344d0-kube-api-access-br4tl\") pod \"9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.045671 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.490579 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq"] Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.598762 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" event={"ID":"71a5370b-ba54-41ae-8e74-38a191f344d0","Type":"ContainerStarted","Data":"f6e916f859486cef5b077fdcaa4cbd6645864c08acd5671e1c8c7cafc579742f"} Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.715616 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw"] Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.718413 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.727705 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw"] Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.771925 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.772013 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.772249 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v575h\" (UniqueName: \"kubernetes.io/projected/11e5d606-14c5-459f-9549-f7ba6c5a698b-kube-api-access-v575h\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.874512 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.874683 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v575h\" (UniqueName: \"kubernetes.io/projected/11e5d606-14c5-459f-9549-f7ba6c5a698b-kube-api-access-v575h\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.874750 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.875239 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-util\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.875571 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-bundle\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:41 crc kubenswrapper[4956]: I1126 17:09:41.904143 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v575h\" (UniqueName: \"kubernetes.io/projected/11e5d606-14c5-459f-9549-f7ba6c5a698b-kube-api-access-v575h\") pod \"87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:42 crc kubenswrapper[4956]: I1126 17:09:42.074224 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:42 crc kubenswrapper[4956]: I1126 17:09:42.454752 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw"] Nov 26 17:09:42 crc kubenswrapper[4956]: W1126 17:09:42.462627 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11e5d606_14c5_459f_9549_f7ba6c5a698b.slice/crio-a37acc6f23e68f16e263d6939c037ee57f9fbbb4f0a5b3a4cb72b25f8cdfe841 WatchSource:0}: Error finding container a37acc6f23e68f16e263d6939c037ee57f9fbbb4f0a5b3a4cb72b25f8cdfe841: Status 404 returned error can't find the container with id a37acc6f23e68f16e263d6939c037ee57f9fbbb4f0a5b3a4cb72b25f8cdfe841 Nov 26 17:09:42 crc kubenswrapper[4956]: I1126 17:09:42.609576 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" event={"ID":"11e5d606-14c5-459f-9549-f7ba6c5a698b","Type":"ContainerStarted","Data":"a37acc6f23e68f16e263d6939c037ee57f9fbbb4f0a5b3a4cb72b25f8cdfe841"} Nov 26 17:09:42 crc kubenswrapper[4956]: I1126 17:09:42.612583 4956 generic.go:334] "Generic (PLEG): container finished" podID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerID="d19a861b6501216812fdbb4b3b7991d014fc8e494cca680710c4f87598b64dc3" exitCode=0 Nov 26 17:09:42 crc kubenswrapper[4956]: I1126 17:09:42.612743 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" event={"ID":"71a5370b-ba54-41ae-8e74-38a191f344d0","Type":"ContainerDied","Data":"d19a861b6501216812fdbb4b3b7991d014fc8e494cca680710c4f87598b64dc3"} Nov 26 17:09:43 crc kubenswrapper[4956]: I1126 17:09:43.628237 4956 generic.go:334] "Generic (PLEG): container finished" podID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerID="1dfd369954b7d6a311375f97ad01eb45482ed8224abe05fd00988028fde39fff" exitCode=0 Nov 26 17:09:43 crc kubenswrapper[4956]: I1126 17:09:43.628320 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" event={"ID":"11e5d606-14c5-459f-9549-f7ba6c5a698b","Type":"ContainerDied","Data":"1dfd369954b7d6a311375f97ad01eb45482ed8224abe05fd00988028fde39fff"} Nov 26 17:09:44 crc kubenswrapper[4956]: I1126 17:09:44.639989 4956 generic.go:334] "Generic (PLEG): container finished" podID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerID="99f773bbcf9fbab449af13802ede65964207e2091224a6c949b5e7147457b870" exitCode=0 Nov 26 17:09:44 crc kubenswrapper[4956]: I1126 17:09:44.640150 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" event={"ID":"71a5370b-ba54-41ae-8e74-38a191f344d0","Type":"ContainerDied","Data":"99f773bbcf9fbab449af13802ede65964207e2091224a6c949b5e7147457b870"} Nov 26 17:09:45 crc kubenswrapper[4956]: I1126 17:09:45.657191 4956 generic.go:334] "Generic (PLEG): container finished" podID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerID="f2d205e45c276c50635146b96afa3bb271000c3f842199eb5e86c96205c991e0" exitCode=0 Nov 26 17:09:45 crc kubenswrapper[4956]: I1126 17:09:45.657315 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" event={"ID":"71a5370b-ba54-41ae-8e74-38a191f344d0","Type":"ContainerDied","Data":"f2d205e45c276c50635146b96afa3bb271000c3f842199eb5e86c96205c991e0"} Nov 26 17:09:45 crc kubenswrapper[4956]: I1126 17:09:45.661491 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" event={"ID":"11e5d606-14c5-459f-9549-f7ba6c5a698b","Type":"ContainerStarted","Data":"861fdb9a8c21b24243e996d5bdc641822b5a8b81a4b4ab08b5134022b15147a3"} Nov 26 17:09:46 crc kubenswrapper[4956]: I1126 17:09:46.916776 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.051556 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-br4tl\" (UniqueName: \"kubernetes.io/projected/71a5370b-ba54-41ae-8e74-38a191f344d0-kube-api-access-br4tl\") pod \"71a5370b-ba54-41ae-8e74-38a191f344d0\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.051654 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-bundle\") pod \"71a5370b-ba54-41ae-8e74-38a191f344d0\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.051724 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-util\") pod \"71a5370b-ba54-41ae-8e74-38a191f344d0\" (UID: \"71a5370b-ba54-41ae-8e74-38a191f344d0\") " Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.052718 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-bundle" (OuterVolumeSpecName: "bundle") pod "71a5370b-ba54-41ae-8e74-38a191f344d0" (UID: "71a5370b-ba54-41ae-8e74-38a191f344d0"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.064247 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/71a5370b-ba54-41ae-8e74-38a191f344d0-kube-api-access-br4tl" (OuterVolumeSpecName: "kube-api-access-br4tl") pod "71a5370b-ba54-41ae-8e74-38a191f344d0" (UID: "71a5370b-ba54-41ae-8e74-38a191f344d0"). InnerVolumeSpecName "kube-api-access-br4tl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.154007 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-br4tl\" (UniqueName: \"kubernetes.io/projected/71a5370b-ba54-41ae-8e74-38a191f344d0-kube-api-access-br4tl\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.154056 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.316432 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-util" (OuterVolumeSpecName: "util") pod "71a5370b-ba54-41ae-8e74-38a191f344d0" (UID: "71a5370b-ba54-41ae-8e74-38a191f344d0"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.357443 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/71a5370b-ba54-41ae-8e74-38a191f344d0-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.679393 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" event={"ID":"71a5370b-ba54-41ae-8e74-38a191f344d0","Type":"ContainerDied","Data":"f6e916f859486cef5b077fdcaa4cbd6645864c08acd5671e1c8c7cafc579742f"} Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.679449 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.679474 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f6e916f859486cef5b077fdcaa4cbd6645864c08acd5671e1c8c7cafc579742f" Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.682266 4956 generic.go:334] "Generic (PLEG): container finished" podID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerID="861fdb9a8c21b24243e996d5bdc641822b5a8b81a4b4ab08b5134022b15147a3" exitCode=0 Nov 26 17:09:47 crc kubenswrapper[4956]: I1126 17:09:47.682307 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" event={"ID":"11e5d606-14c5-459f-9549-f7ba6c5a698b","Type":"ContainerDied","Data":"861fdb9a8c21b24243e996d5bdc641822b5a8b81a4b4ab08b5134022b15147a3"} Nov 26 17:09:48 crc kubenswrapper[4956]: I1126 17:09:48.694800 4956 generic.go:334] "Generic (PLEG): container finished" podID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerID="81cb78d5e289d4c3c268ee8e1914824154d330b8271224afd64ebbbb84622cdb" exitCode=0 Nov 26 17:09:48 crc kubenswrapper[4956]: I1126 17:09:48.694912 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" event={"ID":"11e5d606-14c5-459f-9549-f7ba6c5a698b","Type":"ContainerDied","Data":"81cb78d5e289d4c3c268ee8e1914824154d330b8271224afd64ebbbb84622cdb"} Nov 26 17:09:49 crc kubenswrapper[4956]: I1126 17:09:49.980190 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.037711 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v575h\" (UniqueName: \"kubernetes.io/projected/11e5d606-14c5-459f-9549-f7ba6c5a698b-kube-api-access-v575h\") pod \"11e5d606-14c5-459f-9549-f7ba6c5a698b\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.037811 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-util\") pod \"11e5d606-14c5-459f-9549-f7ba6c5a698b\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.037853 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-bundle\") pod \"11e5d606-14c5-459f-9549-f7ba6c5a698b\" (UID: \"11e5d606-14c5-459f-9549-f7ba6c5a698b\") " Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.039450 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-bundle" (OuterVolumeSpecName: "bundle") pod "11e5d606-14c5-459f-9549-f7ba6c5a698b" (UID: "11e5d606-14c5-459f-9549-f7ba6c5a698b"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.045660 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11e5d606-14c5-459f-9549-f7ba6c5a698b-kube-api-access-v575h" (OuterVolumeSpecName: "kube-api-access-v575h") pod "11e5d606-14c5-459f-9549-f7ba6c5a698b" (UID: "11e5d606-14c5-459f-9549-f7ba6c5a698b"). InnerVolumeSpecName "kube-api-access-v575h". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.057816 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-util" (OuterVolumeSpecName: "util") pod "11e5d606-14c5-459f-9549-f7ba6c5a698b" (UID: "11e5d606-14c5-459f-9549-f7ba6c5a698b"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.139103 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.139194 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/11e5d606-14c5-459f-9549-f7ba6c5a698b-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.139206 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v575h\" (UniqueName: \"kubernetes.io/projected/11e5d606-14c5-459f-9549-f7ba6c5a698b-kube-api-access-v575h\") on node \"crc\" DevicePath \"\"" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.716345 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" event={"ID":"11e5d606-14c5-459f-9549-f7ba6c5a698b","Type":"ContainerDied","Data":"a37acc6f23e68f16e263d6939c037ee57f9fbbb4f0a5b3a4cb72b25f8cdfe841"} Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.716396 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a37acc6f23e68f16e263d6939c037ee57f9fbbb4f0a5b3a4cb72b25f8cdfe841" Nov 26 17:09:50 crc kubenswrapper[4956]: I1126 17:09:50.716486 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.292713 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2"] Nov 26 17:09:59 crc kubenswrapper[4956]: E1126 17:09:59.293594 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="pull" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293611 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="pull" Nov 26 17:09:59 crc kubenswrapper[4956]: E1126 17:09:59.293632 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="extract" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293640 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="extract" Nov 26 17:09:59 crc kubenswrapper[4956]: E1126 17:09:59.293651 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="extract" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293662 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="extract" Nov 26 17:09:59 crc kubenswrapper[4956]: E1126 17:09:59.293680 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="util" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293688 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="util" Nov 26 17:09:59 crc kubenswrapper[4956]: E1126 17:09:59.293710 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="pull" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293717 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="pull" Nov 26 17:09:59 crc kubenswrapper[4956]: E1126 17:09:59.293729 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="util" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293737 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="util" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293905 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="11e5d606-14c5-459f-9549-f7ba6c5a698b" containerName="extract" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.293927 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="71a5370b-ba54-41ae-8e74-38a191f344d0" containerName="extract" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.294509 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.297081 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.297150 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-lvklx" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.313805 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2"] Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.373455 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pq95f\" (UniqueName: \"kubernetes.io/projected/5adb08fd-ab2a-4d2d-afab-4b605419b741-kube-api-access-pq95f\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.373535 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5adb08fd-ab2a-4d2d-afab-4b605419b741-webhook-cert\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.373668 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5adb08fd-ab2a-4d2d-afab-4b605419b741-apiservice-cert\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.474950 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5adb08fd-ab2a-4d2d-afab-4b605419b741-webhook-cert\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.475153 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5adb08fd-ab2a-4d2d-afab-4b605419b741-apiservice-cert\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.475266 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pq95f\" (UniqueName: \"kubernetes.io/projected/5adb08fd-ab2a-4d2d-afab-4b605419b741-kube-api-access-pq95f\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.482772 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/5adb08fd-ab2a-4d2d-afab-4b605419b741-apiservice-cert\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.483787 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/5adb08fd-ab2a-4d2d-afab-4b605419b741-webhook-cert\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.498439 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pq95f\" (UniqueName: \"kubernetes.io/projected/5adb08fd-ab2a-4d2d-afab-4b605419b741-kube-api-access-pq95f\") pod \"horizon-operator-controller-manager-dd7556475-2wzb2\" (UID: \"5adb08fd-ab2a-4d2d-afab-4b605419b741\") " pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:09:59 crc kubenswrapper[4956]: I1126 17:09:59.614288 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:10:00 crc kubenswrapper[4956]: I1126 17:10:00.071349 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2"] Nov 26 17:10:00 crc kubenswrapper[4956]: I1126 17:10:00.801637 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" event={"ID":"5adb08fd-ab2a-4d2d-afab-4b605419b741","Type":"ContainerStarted","Data":"37ee0f284ea2a31297fbd5ebc70b0ebc732c68d6745282dc514d07f2d33beedd"} Nov 26 17:10:03 crc kubenswrapper[4956]: I1126 17:10:03.835685 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" event={"ID":"5adb08fd-ab2a-4d2d-afab-4b605419b741","Type":"ContainerStarted","Data":"2c3e3e9bc5551b703165baa12edde7bfcf326212884f441003436889505bb0a2"} Nov 26 17:10:03 crc kubenswrapper[4956]: I1126 17:10:03.836675 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:10:03 crc kubenswrapper[4956]: I1126 17:10:03.855423 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" podStartSLOduration=2.142662768 podStartE2EDuration="4.855397437s" podCreationTimestamp="2025-11-26 17:09:59 +0000 UTC" firstStartedPulling="2025-11-26 17:10:00.081184198 +0000 UTC m=+1045.777144750" lastFinishedPulling="2025-11-26 17:10:02.793918857 +0000 UTC m=+1048.489879419" observedRunningTime="2025-11-26 17:10:03.853086061 +0000 UTC m=+1049.549046623" watchObservedRunningTime="2025-11-26 17:10:03.855397437 +0000 UTC m=+1049.551357989" Nov 26 17:10:05 crc kubenswrapper[4956]: I1126 17:10:05.502978 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/keystone-6c49f4fcbc-tchbn" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.086808 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb"] Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.089428 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.092182 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lvvl8" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.092613 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.106533 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb"] Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.127091 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1090fc6e-22d8-4fb2-9296-46edde4e1af5-apiservice-cert\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.127169 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1090fc6e-22d8-4fb2-9296-46edde4e1af5-webhook-cert\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.228537 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bjgbq\" (UniqueName: \"kubernetes.io/projected/1090fc6e-22d8-4fb2-9296-46edde4e1af5-kube-api-access-bjgbq\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.228631 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1090fc6e-22d8-4fb2-9296-46edde4e1af5-apiservice-cert\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.229206 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1090fc6e-22d8-4fb2-9296-46edde4e1af5-webhook-cert\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.238130 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/1090fc6e-22d8-4fb2-9296-46edde4e1af5-webhook-cert\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.238130 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/1090fc6e-22d8-4fb2-9296-46edde4e1af5-apiservice-cert\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.331113 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bjgbq\" (UniqueName: \"kubernetes.io/projected/1090fc6e-22d8-4fb2-9296-46edde4e1af5-kube-api-access-bjgbq\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.352797 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bjgbq\" (UniqueName: \"kubernetes.io/projected/1090fc6e-22d8-4fb2-9296-46edde4e1af5-kube-api-access-bjgbq\") pod \"swift-operator-controller-manager-f659d5cfc-2plpb\" (UID: \"1090fc6e-22d8-4fb2-9296-46edde4e1af5\") " pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.425385 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:08 crc kubenswrapper[4956]: I1126 17:10:08.860237 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb"] Nov 26 17:10:09 crc kubenswrapper[4956]: I1126 17:10:09.620962 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:10:09 crc kubenswrapper[4956]: I1126 17:10:09.882110 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" event={"ID":"1090fc6e-22d8-4fb2-9296-46edde4e1af5","Type":"ContainerStarted","Data":"917f02a780fb080f5c368a293905fcaa0beb3f05ae3178e7cef8b5c16a1f8dc9"} Nov 26 17:10:11 crc kubenswrapper[4956]: I1126 17:10:11.899669 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" event={"ID":"1090fc6e-22d8-4fb2-9296-46edde4e1af5","Type":"ContainerStarted","Data":"c5eaa44ed20bf03ec8473cda3e7348f1ca69b6e79c0fc8b463d29691c017b92a"} Nov 26 17:10:11 crc kubenswrapper[4956]: I1126 17:10:11.900638 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:11 crc kubenswrapper[4956]: I1126 17:10:11.934553 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" podStartSLOduration=1.48213973 podStartE2EDuration="3.934518086s" podCreationTimestamp="2025-11-26 17:10:08 +0000 UTC" firstStartedPulling="2025-11-26 17:10:08.868225802 +0000 UTC m=+1054.564186354" lastFinishedPulling="2025-11-26 17:10:11.320604158 +0000 UTC m=+1057.016564710" observedRunningTime="2025-11-26 17:10:11.930538043 +0000 UTC m=+1057.626498595" watchObservedRunningTime="2025-11-26 17:10:11.934518086 +0000 UTC m=+1057.630478638" Nov 26 17:10:18 crc kubenswrapper[4956]: I1126 17:10:18.433662 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.135944 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.142713 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.145681 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.146145 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.146231 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.149508 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-zvkwr" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.171123 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.184198 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.184270 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.184307 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/26554e9d-3393-425b-bc71-07e66397ca3b-cache\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.184343 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q947z\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-kube-api-access-q947z\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.184372 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/26554e9d-3393-425b-bc71-07e66397ca3b-lock\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.286383 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.286465 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.286511 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/26554e9d-3393-425b-bc71-07e66397ca3b-cache\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.286561 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q947z\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-kube-api-access-q947z\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.287160 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") device mount path \"/mnt/openstack/pv08\"" pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: E1126 17:10:24.286654 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:24 crc kubenswrapper[4956]: E1126 17:10:24.287260 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 26 17:10:24 crc kubenswrapper[4956]: E1126 17:10:24.287343 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift podName:26554e9d-3393-425b-bc71-07e66397ca3b nodeName:}" failed. No retries permitted until 2025-11-26 17:10:24.787310698 +0000 UTC m=+1070.483271250 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift") pod "swift-storage-0" (UID: "26554e9d-3393-425b-bc71-07e66397ca3b") : configmap "swift-ring-files" not found Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.287253 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/26554e9d-3393-425b-bc71-07e66397ca3b-cache\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.287205 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/26554e9d-3393-425b-bc71-07e66397ca3b-lock\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.287733 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/26554e9d-3393-425b-bc71-07e66397ca3b-lock\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.316740 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.323914 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q947z\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-kube-api-access-q947z\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: I1126 17:10:24.795055 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:24 crc kubenswrapper[4956]: E1126 17:10:24.795390 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:24 crc kubenswrapper[4956]: E1126 17:10:24.795456 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 26 17:10:24 crc kubenswrapper[4956]: E1126 17:10:24.795557 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift podName:26554e9d-3393-425b-bc71-07e66397ca3b nodeName:}" failed. No retries permitted until 2025-11-26 17:10:25.795527589 +0000 UTC m=+1071.491488181 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift") pod "swift-storage-0" (UID: "26554e9d-3393-425b-bc71-07e66397ca3b") : configmap "swift-ring-files" not found Nov 26 17:10:25 crc kubenswrapper[4956]: I1126 17:10:25.811798 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:25 crc kubenswrapper[4956]: E1126 17:10:25.812296 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:25 crc kubenswrapper[4956]: E1126 17:10:25.812406 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 26 17:10:25 crc kubenswrapper[4956]: E1126 17:10:25.812452 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift podName:26554e9d-3393-425b-bc71-07e66397ca3b nodeName:}" failed. No retries permitted until 2025-11-26 17:10:27.812438485 +0000 UTC m=+1073.508399037 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift") pod "swift-storage-0" (UID: "26554e9d-3393-425b-bc71-07e66397ca3b") : configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.072464 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-index-9pjcm"] Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.073889 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.076683 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-j8svs" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.085026 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-9pjcm"] Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.119251 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8th9w\" (UniqueName: \"kubernetes.io/projected/e9e2f38e-3b35-4c00-ab0b-36602b28a4f0-kube-api-access-8th9w\") pod \"glance-operator-index-9pjcm\" (UID: \"e9e2f38e-3b35-4c00-ab0b-36602b28a4f0\") " pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.176690 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd"] Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.178180 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.180393 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.196683 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd"] Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.220598 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rqv7m\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-kube-api-access-rqv7m\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.220664 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f91290f1-6c5a-4199-806d-ab204e05814e-run-httpd\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.220703 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.220755 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f91290f1-6c5a-4199-806d-ab204e05814e-log-httpd\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.220814 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8th9w\" (UniqueName: \"kubernetes.io/projected/e9e2f38e-3b35-4c00-ab0b-36602b28a4f0-kube-api-access-8th9w\") pod \"glance-operator-index-9pjcm\" (UID: \"e9e2f38e-3b35-4c00-ab0b-36602b28a4f0\") " pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.220850 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91290f1-6c5a-4199-806d-ab204e05814e-config-data\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.244599 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8th9w\" (UniqueName: \"kubernetes.io/projected/e9e2f38e-3b35-4c00-ab0b-36602b28a4f0-kube-api-access-8th9w\") pod \"glance-operator-index-9pjcm\" (UID: \"e9e2f38e-3b35-4c00-ab0b-36602b28a4f0\") " pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.322822 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.322935 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f91290f1-6c5a-4199-806d-ab204e05814e-log-httpd\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.322992 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91290f1-6c5a-4199-806d-ab204e05814e-config-data\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.323048 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rqv7m\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-kube-api-access-rqv7m\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.323077 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f91290f1-6c5a-4199-806d-ab204e05814e-run-httpd\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: E1126 17:10:26.323133 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: E1126 17:10:26.323179 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd: configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: E1126 17:10:26.323263 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift podName:f91290f1-6c5a-4199-806d-ab204e05814e nodeName:}" failed. No retries permitted until 2025-11-26 17:10:26.82322756 +0000 UTC m=+1072.519188282 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift") pod "swift-proxy-6bd58cfcf7-bj6nd" (UID: "f91290f1-6c5a-4199-806d-ab204e05814e") : configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.323608 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f91290f1-6c5a-4199-806d-ab204e05814e-run-httpd\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.324101 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/f91290f1-6c5a-4199-806d-ab204e05814e-log-httpd\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.329382 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f91290f1-6c5a-4199-806d-ab204e05814e-config-data\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.340530 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rqv7m\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-kube-api-access-rqv7m\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.397616 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.832685 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:26 crc kubenswrapper[4956]: E1126 17:10:26.832898 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: E1126 17:10:26.833199 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd: configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: E1126 17:10:26.833264 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift podName:f91290f1-6c5a-4199-806d-ab204e05814e nodeName:}" failed. No retries permitted until 2025-11-26 17:10:27.833243293 +0000 UTC m=+1073.529203845 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift") pod "swift-proxy-6bd58cfcf7-bj6nd" (UID: "f91290f1-6c5a-4199-806d-ab204e05814e") : configmap "swift-ring-files" not found Nov 26 17:10:26 crc kubenswrapper[4956]: I1126 17:10:26.943471 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-index-9pjcm"] Nov 26 17:10:27 crc kubenswrapper[4956]: I1126 17:10:27.033849 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-9pjcm" event={"ID":"e9e2f38e-3b35-4c00-ab0b-36602b28a4f0","Type":"ContainerStarted","Data":"00841db9ef7b5f69f727a4b490c042e6ad00aae9bc8bba14a18eff5366a80b05"} Nov 26 17:10:27 crc kubenswrapper[4956]: I1126 17:10:27.851006 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:27 crc kubenswrapper[4956]: I1126 17:10:27.851439 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:27 crc kubenswrapper[4956]: E1126 17:10:27.851210 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:27 crc kubenswrapper[4956]: E1126 17:10:27.851494 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd: configmap "swift-ring-files" not found Nov 26 17:10:27 crc kubenswrapper[4956]: E1126 17:10:27.851555 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift podName:f91290f1-6c5a-4199-806d-ab204e05814e nodeName:}" failed. No retries permitted until 2025-11-26 17:10:29.85153596 +0000 UTC m=+1075.547496512 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift") pod "swift-proxy-6bd58cfcf7-bj6nd" (UID: "f91290f1-6c5a-4199-806d-ab204e05814e") : configmap "swift-ring-files" not found Nov 26 17:10:27 crc kubenswrapper[4956]: E1126 17:10:27.851561 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:27 crc kubenswrapper[4956]: E1126 17:10:27.851572 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 26 17:10:27 crc kubenswrapper[4956]: E1126 17:10:27.851600 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift podName:26554e9d-3393-425b-bc71-07e66397ca3b nodeName:}" failed. No retries permitted until 2025-11-26 17:10:31.851590341 +0000 UTC m=+1077.547550893 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift") pod "swift-storage-0" (UID: "26554e9d-3393-425b-bc71-07e66397ca3b") : configmap "swift-ring-files" not found Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.194974 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-r72t9"] Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.196301 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.202906 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-scripts" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.203282 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-config-data" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.210044 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-r72t9"] Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.258265 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-dispersionconf\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.258376 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-etc-swift\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.258417 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-swiftconf\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.258458 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hcwfz\" (UniqueName: \"kubernetes.io/projected/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-kube-api-access-hcwfz\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.258532 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-ring-data-devices\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.258573 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-scripts\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.359780 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-dispersionconf\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.359891 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-etc-swift\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.359922 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-swiftconf\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.359949 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hcwfz\" (UniqueName: \"kubernetes.io/projected/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-kube-api-access-hcwfz\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.359988 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-ring-data-devices\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.360016 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-scripts\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.360557 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-etc-swift\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.361056 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-scripts\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.361411 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-ring-data-devices\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.372275 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-dispersionconf\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.382645 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-swiftconf\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.384033 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hcwfz\" (UniqueName: \"kubernetes.io/projected/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-kube-api-access-hcwfz\") pod \"swift-ring-rebalance-r72t9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:28 crc kubenswrapper[4956]: I1126 17:10:28.516604 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:29 crc kubenswrapper[4956]: I1126 17:10:29.019095 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-ring-rebalance-r72t9"] Nov 26 17:10:29 crc kubenswrapper[4956]: I1126 17:10:29.889709 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:29 crc kubenswrapper[4956]: E1126 17:10:29.890210 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:29 crc kubenswrapper[4956]: E1126 17:10:29.891553 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd: configmap "swift-ring-files" not found Nov 26 17:10:29 crc kubenswrapper[4956]: E1126 17:10:29.891693 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift podName:f91290f1-6c5a-4199-806d-ab204e05814e nodeName:}" failed. No retries permitted until 2025-11-26 17:10:33.891659063 +0000 UTC m=+1079.587619615 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift") pod "swift-proxy-6bd58cfcf7-bj6nd" (UID: "f91290f1-6c5a-4199-806d-ab204e05814e") : configmap "swift-ring-files" not found Nov 26 17:10:30 crc kubenswrapper[4956]: I1126 17:10:30.059647 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" event={"ID":"83a3f8ba-c224-4656-b7a6-8ac28c1804f9","Type":"ContainerStarted","Data":"7763ab7dfe82b7d057b5fce4e0a12cb257eccdef121dbad019e3ffa59f9b4efb"} Nov 26 17:10:31 crc kubenswrapper[4956]: I1126 17:10:31.076767 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-index-9pjcm" event={"ID":"e9e2f38e-3b35-4c00-ab0b-36602b28a4f0","Type":"ContainerStarted","Data":"2bc00f4773284f61b58268119967e94d9d808213af8d7a141f6420db0ae43eca"} Nov 26 17:10:31 crc kubenswrapper[4956]: I1126 17:10:31.090848 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-index-9pjcm" podStartSLOduration=2.097809762 podStartE2EDuration="5.090824834s" podCreationTimestamp="2025-11-26 17:10:26 +0000 UTC" firstStartedPulling="2025-11-26 17:10:26.956462152 +0000 UTC m=+1072.652422704" lastFinishedPulling="2025-11-26 17:10:29.949477224 +0000 UTC m=+1075.645437776" observedRunningTime="2025-11-26 17:10:31.090440893 +0000 UTC m=+1076.786401445" watchObservedRunningTime="2025-11-26 17:10:31.090824834 +0000 UTC m=+1076.786785386" Nov 26 17:10:31 crc kubenswrapper[4956]: I1126 17:10:31.924003 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:31 crc kubenswrapper[4956]: E1126 17:10:31.924303 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:31 crc kubenswrapper[4956]: E1126 17:10:31.924732 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 26 17:10:31 crc kubenswrapper[4956]: E1126 17:10:31.924889 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift podName:26554e9d-3393-425b-bc71-07e66397ca3b nodeName:}" failed. No retries permitted until 2025-11-26 17:10:39.924845909 +0000 UTC m=+1085.620806461 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift") pod "swift-storage-0" (UID: "26554e9d-3393-425b-bc71-07e66397ca3b") : configmap "swift-ring-files" not found Nov 26 17:10:33 crc kubenswrapper[4956]: I1126 17:10:33.965157 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:33 crc kubenswrapper[4956]: E1126 17:10:33.965350 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:33 crc kubenswrapper[4956]: E1126 17:10:33.965413 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd: configmap "swift-ring-files" not found Nov 26 17:10:33 crc kubenswrapper[4956]: E1126 17:10:33.965488 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift podName:f91290f1-6c5a-4199-806d-ab204e05814e nodeName:}" failed. No retries permitted until 2025-11-26 17:10:41.965466056 +0000 UTC m=+1087.661426608 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift") pod "swift-proxy-6bd58cfcf7-bj6nd" (UID: "f91290f1-6c5a-4199-806d-ab204e05814e") : configmap "swift-ring-files" not found Nov 26 17:10:36 crc kubenswrapper[4956]: I1126 17:10:36.398644 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:36 crc kubenswrapper[4956]: I1126 17:10:36.399269 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:36 crc kubenswrapper[4956]: I1126 17:10:36.430498 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:37 crc kubenswrapper[4956]: I1126 17:10:37.177413 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-index-9pjcm" Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.918036 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l"] Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.922580 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.925620 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-j97nd" Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.926543 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l"] Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.976126 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-util\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.976177 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-bundle\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:38 crc kubenswrapper[4956]: I1126 17:10:38.976203 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-npb46\" (UniqueName: \"kubernetes.io/projected/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-kube-api-access-npb46\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.077140 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-util\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.077200 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-bundle\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.077222 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-npb46\" (UniqueName: \"kubernetes.io/projected/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-kube-api-access-npb46\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.078072 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-bundle\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.078081 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-util\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.101762 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-npb46\" (UniqueName: \"kubernetes.io/projected/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-kube-api-access-npb46\") pod \"10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.173200 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" event={"ID":"83a3f8ba-c224-4656-b7a6-8ac28c1804f9","Type":"ContainerStarted","Data":"9a8ac2d199b8d1b468b78df49f6f14d848e3a7e7e7e97759e71131efbf636c7d"} Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.278584 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.765795 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l"] Nov 26 17:10:39 crc kubenswrapper[4956]: I1126 17:10:39.996286 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:39 crc kubenswrapper[4956]: E1126 17:10:39.996676 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:39 crc kubenswrapper[4956]: E1126 17:10:39.996917 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-storage-0: configmap "swift-ring-files" not found Nov 26 17:10:39 crc kubenswrapper[4956]: E1126 17:10:39.998043 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift podName:26554e9d-3393-425b-bc71-07e66397ca3b nodeName:}" failed. No retries permitted until 2025-11-26 17:10:55.996966539 +0000 UTC m=+1101.692927101 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift") pod "swift-storage-0" (UID: "26554e9d-3393-425b-bc71-07e66397ca3b") : configmap "swift-ring-files" not found Nov 26 17:10:40 crc kubenswrapper[4956]: I1126 17:10:40.185024 4956 generic.go:334] "Generic (PLEG): container finished" podID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerID="93a653511115bf8c56c3c09201586a0a20146181f79a285c90e913072391330b" exitCode=0 Nov 26 17:10:40 crc kubenswrapper[4956]: I1126 17:10:40.185144 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" event={"ID":"c6bd841b-0f31-4fd1-94c2-d776ffe4582c","Type":"ContainerDied","Data":"93a653511115bf8c56c3c09201586a0a20146181f79a285c90e913072391330b"} Nov 26 17:10:40 crc kubenswrapper[4956]: I1126 17:10:40.185221 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" event={"ID":"c6bd841b-0f31-4fd1-94c2-d776ffe4582c","Type":"ContainerStarted","Data":"fc48314b79e24f3d888d61929dd9e902bfc651677464343561e6716c58336b16"} Nov 26 17:10:40 crc kubenswrapper[4956]: I1126 17:10:40.286985 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" podStartSLOduration=3.262162955 podStartE2EDuration="12.286930878s" podCreationTimestamp="2025-11-26 17:10:28 +0000 UTC" firstStartedPulling="2025-11-26 17:10:29.779983374 +0000 UTC m=+1075.475943916" lastFinishedPulling="2025-11-26 17:10:38.804751277 +0000 UTC m=+1084.500711839" observedRunningTime="2025-11-26 17:10:40.280926107 +0000 UTC m=+1085.976886669" watchObservedRunningTime="2025-11-26 17:10:40.286930878 +0000 UTC m=+1085.982891430" Nov 26 17:10:42 crc kubenswrapper[4956]: I1126 17:10:42.031701 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:42 crc kubenswrapper[4956]: E1126 17:10:42.032010 4956 projected.go:288] Couldn't get configMap glance-kuttl-tests/swift-ring-files: configmap "swift-ring-files" not found Nov 26 17:10:42 crc kubenswrapper[4956]: E1126 17:10:42.032748 4956 projected.go:194] Error preparing data for projected volume etc-swift for pod glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd: configmap "swift-ring-files" not found Nov 26 17:10:42 crc kubenswrapper[4956]: E1126 17:10:42.032842 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift podName:f91290f1-6c5a-4199-806d-ab204e05814e nodeName:}" failed. No retries permitted until 2025-11-26 17:10:58.032817401 +0000 UTC m=+1103.728777953 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift") pod "swift-proxy-6bd58cfcf7-bj6nd" (UID: "f91290f1-6c5a-4199-806d-ab204e05814e") : configmap "swift-ring-files" not found Nov 26 17:10:42 crc kubenswrapper[4956]: I1126 17:10:42.203103 4956 generic.go:334] "Generic (PLEG): container finished" podID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerID="e1ad4c188ec02fdb1553b7dc84c013ae05bb34b3b9dc9279bdb6335a5c2c4bdc" exitCode=0 Nov 26 17:10:42 crc kubenswrapper[4956]: I1126 17:10:42.203159 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" event={"ID":"c6bd841b-0f31-4fd1-94c2-d776ffe4582c","Type":"ContainerDied","Data":"e1ad4c188ec02fdb1553b7dc84c013ae05bb34b3b9dc9279bdb6335a5c2c4bdc"} Nov 26 17:10:43 crc kubenswrapper[4956]: I1126 17:10:43.213968 4956 generic.go:334] "Generic (PLEG): container finished" podID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerID="00653cc1493ad7dc37937d01d40f2e7be7d772da88f3a1c560c1117373e0c777" exitCode=0 Nov 26 17:10:43 crc kubenswrapper[4956]: I1126 17:10:43.214035 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" event={"ID":"c6bd841b-0f31-4fd1-94c2-d776ffe4582c","Type":"ContainerDied","Data":"00653cc1493ad7dc37937d01d40f2e7be7d772da88f3a1c560c1117373e0c777"} Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.568137 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.678463 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-npb46\" (UniqueName: \"kubernetes.io/projected/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-kube-api-access-npb46\") pod \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.678696 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-bundle\") pod \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.678934 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-util\") pod \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\" (UID: \"c6bd841b-0f31-4fd1-94c2-d776ffe4582c\") " Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.680080 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-bundle" (OuterVolumeSpecName: "bundle") pod "c6bd841b-0f31-4fd1-94c2-d776ffe4582c" (UID: "c6bd841b-0f31-4fd1-94c2-d776ffe4582c"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.686307 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-kube-api-access-npb46" (OuterVolumeSpecName: "kube-api-access-npb46") pod "c6bd841b-0f31-4fd1-94c2-d776ffe4582c" (UID: "c6bd841b-0f31-4fd1-94c2-d776ffe4582c"). InnerVolumeSpecName "kube-api-access-npb46". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.693749 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-util" (OuterVolumeSpecName: "util") pod "c6bd841b-0f31-4fd1-94c2-d776ffe4582c" (UID: "c6bd841b-0f31-4fd1-94c2-d776ffe4582c"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.780690 4956 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.780737 4956 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-util\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:44 crc kubenswrapper[4956]: I1126 17:10:44.780747 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-npb46\" (UniqueName: \"kubernetes.io/projected/c6bd841b-0f31-4fd1-94c2-d776ffe4582c-kube-api-access-npb46\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:45 crc kubenswrapper[4956]: I1126 17:10:45.232347 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" event={"ID":"c6bd841b-0f31-4fd1-94c2-d776ffe4582c","Type":"ContainerDied","Data":"fc48314b79e24f3d888d61929dd9e902bfc651677464343561e6716c58336b16"} Nov 26 17:10:45 crc kubenswrapper[4956]: I1126 17:10:45.232910 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc48314b79e24f3d888d61929dd9e902bfc651677464343561e6716c58336b16" Nov 26 17:10:45 crc kubenswrapper[4956]: I1126 17:10:45.232395 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l" Nov 26 17:10:47 crc kubenswrapper[4956]: I1126 17:10:47.249937 4956 generic.go:334] "Generic (PLEG): container finished" podID="83a3f8ba-c224-4656-b7a6-8ac28c1804f9" containerID="9a8ac2d199b8d1b468b78df49f6f14d848e3a7e7e7e97759e71131efbf636c7d" exitCode=0 Nov 26 17:10:47 crc kubenswrapper[4956]: I1126 17:10:47.250059 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" event={"ID":"83a3f8ba-c224-4656-b7a6-8ac28c1804f9","Type":"ContainerDied","Data":"9a8ac2d199b8d1b468b78df49f6f14d848e3a7e7e7e97759e71131efbf636c7d"} Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.647385 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.760721 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-etc-swift\") pod \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.760802 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hcwfz\" (UniqueName: \"kubernetes.io/projected/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-kube-api-access-hcwfz\") pod \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.760843 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-dispersionconf\") pod \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.760982 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-ring-data-devices\") pod \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.761017 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-swiftconf\") pod \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.761048 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-scripts\") pod \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\" (UID: \"83a3f8ba-c224-4656-b7a6-8ac28c1804f9\") " Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.761625 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "83a3f8ba-c224-4656-b7a6-8ac28c1804f9" (UID: "83a3f8ba-c224-4656-b7a6-8ac28c1804f9"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.761787 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "83a3f8ba-c224-4656-b7a6-8ac28c1804f9" (UID: "83a3f8ba-c224-4656-b7a6-8ac28c1804f9"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.768470 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-kube-api-access-hcwfz" (OuterVolumeSpecName: "kube-api-access-hcwfz") pod "83a3f8ba-c224-4656-b7a6-8ac28c1804f9" (UID: "83a3f8ba-c224-4656-b7a6-8ac28c1804f9"). InnerVolumeSpecName "kube-api-access-hcwfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.785569 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-scripts" (OuterVolumeSpecName: "scripts") pod "83a3f8ba-c224-4656-b7a6-8ac28c1804f9" (UID: "83a3f8ba-c224-4656-b7a6-8ac28c1804f9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.787329 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "83a3f8ba-c224-4656-b7a6-8ac28c1804f9" (UID: "83a3f8ba-c224-4656-b7a6-8ac28c1804f9"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.789744 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "83a3f8ba-c224-4656-b7a6-8ac28c1804f9" (UID: "83a3f8ba-c224-4656-b7a6-8ac28c1804f9"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.863031 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hcwfz\" (UniqueName: \"kubernetes.io/projected/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-kube-api-access-hcwfz\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.863065 4956 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-dispersionconf\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.863074 4956 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-ring-data-devices\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.863085 4956 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-swiftconf\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.863094 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:48 crc kubenswrapper[4956]: I1126 17:10:48.863109 4956 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/83a3f8ba-c224-4656-b7a6-8ac28c1804f9-etc-swift\") on node \"crc\" DevicePath \"\"" Nov 26 17:10:49 crc kubenswrapper[4956]: I1126 17:10:49.268351 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" event={"ID":"83a3f8ba-c224-4656-b7a6-8ac28c1804f9","Type":"ContainerDied","Data":"7763ab7dfe82b7d057b5fce4e0a12cb257eccdef121dbad019e3ffa59f9b4efb"} Nov 26 17:10:49 crc kubenswrapper[4956]: I1126 17:10:49.268397 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-ring-rebalance-r72t9" Nov 26 17:10:49 crc kubenswrapper[4956]: I1126 17:10:49.268420 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7763ab7dfe82b7d057b5fce4e0a12cb257eccdef121dbad019e3ffa59f9b4efb" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.446446 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6"] Nov 26 17:10:55 crc kubenswrapper[4956]: E1126 17:10:55.447544 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="util" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.447567 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="util" Nov 26 17:10:55 crc kubenswrapper[4956]: E1126 17:10:55.447601 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="pull" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.447609 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="pull" Nov 26 17:10:55 crc kubenswrapper[4956]: E1126 17:10:55.447623 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83a3f8ba-c224-4656-b7a6-8ac28c1804f9" containerName="swift-ring-rebalance" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.447631 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="83a3f8ba-c224-4656-b7a6-8ac28c1804f9" containerName="swift-ring-rebalance" Nov 26 17:10:55 crc kubenswrapper[4956]: E1126 17:10:55.447643 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="extract" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.447651 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="extract" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.447802 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="83a3f8ba-c224-4656-b7a6-8ac28c1804f9" containerName="swift-ring-rebalance" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.447823 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="c6bd841b-0f31-4fd1-94c2-d776ffe4582c" containerName="extract" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.448441 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.453035 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zs4z6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.453408 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.467793 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6"] Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.582683 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d2238f24-43e6-4512-880a-1c3b10ba3844-webhook-cert\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.582742 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d2238f24-43e6-4512-880a-1c3b10ba3844-apiservice-cert\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.582845 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nfb6x\" (UniqueName: \"kubernetes.io/projected/d2238f24-43e6-4512-880a-1c3b10ba3844-kube-api-access-nfb6x\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.684420 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nfb6x\" (UniqueName: \"kubernetes.io/projected/d2238f24-43e6-4512-880a-1c3b10ba3844-kube-api-access-nfb6x\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.684606 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d2238f24-43e6-4512-880a-1c3b10ba3844-webhook-cert\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.684633 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d2238f24-43e6-4512-880a-1c3b10ba3844-apiservice-cert\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.691679 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/d2238f24-43e6-4512-880a-1c3b10ba3844-apiservice-cert\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.691725 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/d2238f24-43e6-4512-880a-1c3b10ba3844-webhook-cert\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.705347 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nfb6x\" (UniqueName: \"kubernetes.io/projected/d2238f24-43e6-4512-880a-1c3b10ba3844-kube-api-access-nfb6x\") pod \"glance-operator-controller-manager-788c9777f6-7qmf6\" (UID: \"d2238f24-43e6-4512-880a-1c3b10ba3844\") " pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:55 crc kubenswrapper[4956]: I1126 17:10:55.811595 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:10:56 crc kubenswrapper[4956]: I1126 17:10:56.090951 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:56 crc kubenswrapper[4956]: I1126 17:10:56.100626 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/26554e9d-3393-425b-bc71-07e66397ca3b-etc-swift\") pod \"swift-storage-0\" (UID: \"26554e9d-3393-425b-bc71-07e66397ca3b\") " pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:56 crc kubenswrapper[4956]: I1126 17:10:56.261534 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-storage-0" Nov 26 17:10:56 crc kubenswrapper[4956]: I1126 17:10:56.285786 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6"] Nov 26 17:10:56 crc kubenswrapper[4956]: I1126 17:10:56.330458 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" event={"ID":"d2238f24-43e6-4512-880a-1c3b10ba3844","Type":"ContainerStarted","Data":"8031ba926029319509ec9991324e38300ad1e3508d9d629900a51faca4cfeb34"} Nov 26 17:10:56 crc kubenswrapper[4956]: I1126 17:10:56.759000 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-storage-0"] Nov 26 17:10:56 crc kubenswrapper[4956]: W1126 17:10:56.771014 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26554e9d_3393_425b_bc71_07e66397ca3b.slice/crio-cb77f59bdb8c645b343e7f06864aff8f169c81d03d9dd56c7a6cc22b8d028d25 WatchSource:0}: Error finding container cb77f59bdb8c645b343e7f06864aff8f169c81d03d9dd56c7a6cc22b8d028d25: Status 404 returned error can't find the container with id cb77f59bdb8c645b343e7f06864aff8f169c81d03d9dd56c7a6cc22b8d028d25 Nov 26 17:10:57 crc kubenswrapper[4956]: I1126 17:10:57.350357 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"cb77f59bdb8c645b343e7f06864aff8f169c81d03d9dd56c7a6cc22b8d028d25"} Nov 26 17:10:58 crc kubenswrapper[4956]: I1126 17:10:58.133781 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:58 crc kubenswrapper[4956]: I1126 17:10:58.141825 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/f91290f1-6c5a-4199-806d-ab204e05814e-etc-swift\") pod \"swift-proxy-6bd58cfcf7-bj6nd\" (UID: \"f91290f1-6c5a-4199-806d-ab204e05814e\") " pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:10:58 crc kubenswrapper[4956]: I1126 17:10:58.295538 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:11:00 crc kubenswrapper[4956]: I1126 17:11:00.211053 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd"] Nov 26 17:11:00 crc kubenswrapper[4956]: I1126 17:11:00.392821 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" event={"ID":"f91290f1-6c5a-4199-806d-ab204e05814e","Type":"ContainerStarted","Data":"064a5bd2f45da872f3311825bb3ffbaa6e69aab9dd1a6ebaaba9996c62715377"} Nov 26 17:11:00 crc kubenswrapper[4956]: I1126 17:11:00.395386 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" event={"ID":"d2238f24-43e6-4512-880a-1c3b10ba3844","Type":"ContainerStarted","Data":"90329b5d7c5468486a729f757cfd442e663d9c08c31c3bb62235099f0dfcf5c5"} Nov 26 17:11:00 crc kubenswrapper[4956]: I1126 17:11:00.395668 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:11:00 crc kubenswrapper[4956]: I1126 17:11:00.417546 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" podStartSLOduration=1.791675651 podStartE2EDuration="5.417525396s" podCreationTimestamp="2025-11-26 17:10:55 +0000 UTC" firstStartedPulling="2025-11-26 17:10:56.292894462 +0000 UTC m=+1101.988855024" lastFinishedPulling="2025-11-26 17:10:59.918744197 +0000 UTC m=+1105.614704769" observedRunningTime="2025-11-26 17:11:00.415317613 +0000 UTC m=+1106.111278175" watchObservedRunningTime="2025-11-26 17:11:00.417525396 +0000 UTC m=+1106.113485948" Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.407237 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"29613e53ec21699fa430e85d65e23d622d457572ad87cef1ec41fe78a79558e1"} Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.407958 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"e4c5c32182265b183ecb0284db29ce00da78645e70bc9cd516b5747816260ab0"} Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.410919 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" event={"ID":"f91290f1-6c5a-4199-806d-ab204e05814e","Type":"ContainerStarted","Data":"410f0888c9c6d4c29aa3c7c9f9d86ac4eea5b2156c416b232857127ce3aa3070"} Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.410945 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" event={"ID":"f91290f1-6c5a-4199-806d-ab204e05814e","Type":"ContainerStarted","Data":"0135a73d6d20f6058674959a67e62f10231c314dbf31d90e8ef487a5b7e7a343"} Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.410964 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.410975 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:11:01 crc kubenswrapper[4956]: I1126 17:11:01.446608 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" podStartSLOduration=35.446561284 podStartE2EDuration="35.446561284s" podCreationTimestamp="2025-11-26 17:10:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:11:01.434829318 +0000 UTC m=+1107.130789890" watchObservedRunningTime="2025-11-26 17:11:01.446561284 +0000 UTC m=+1107.142521836" Nov 26 17:11:02 crc kubenswrapper[4956]: I1126 17:11:02.419197 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"d440e8c93aa41e95686ad05564b6688db3bcac9b5e04f35e8253ca53fd626196"} Nov 26 17:11:02 crc kubenswrapper[4956]: I1126 17:11:02.419629 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"b32f7fe5b01c45e9cf86e4e7240fc1a40a1fe26e1eae3df4e1994b4727899da3"} Nov 26 17:11:05 crc kubenswrapper[4956]: I1126 17:11:05.465421 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"c9a428684cf1c6b6ed1bd2f42c0dcbe1dc4627f13579eee8727984bf6db23216"} Nov 26 17:11:05 crc kubenswrapper[4956]: I1126 17:11:05.815968 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:11:06 crc kubenswrapper[4956]: I1126 17:11:06.479597 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"dfdcee26151198b566aa5e99fe6d264203e84f49409c3341dfe9f35321146bf0"} Nov 26 17:11:07 crc kubenswrapper[4956]: I1126 17:11:07.495340 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"070648626c4d1cc16c49c38258568f2049a7d23c8d6bead5a334731d3585caf3"} Nov 26 17:11:07 crc kubenswrapper[4956]: I1126 17:11:07.495835 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"c66e6b47544e1336fa38194b6eb6742316af8ed3abc247d6c0bd00a264877238"} Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.304893 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.305845 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-nmgcz"] Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.306796 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.310976 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/swift-proxy-6bd58cfcf7-bj6nd" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.335763 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-59e4-account-create-update-wfrc2"] Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.337038 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.341470 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-nmgcz"] Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.341913 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.355171 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfd4958-948b-4f5b-86ee-1c0c5242df17-operator-scripts\") pod \"glance-db-create-nmgcz\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.355284 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1ec54ce-5118-4f19-b962-a4e0100f08e6-operator-scripts\") pod \"glance-59e4-account-create-update-wfrc2\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.355333 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kvvcs\" (UniqueName: \"kubernetes.io/projected/b1ec54ce-5118-4f19-b962-a4e0100f08e6-kube-api-access-kvvcs\") pod \"glance-59e4-account-create-update-wfrc2\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.355357 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rtfhb\" (UniqueName: \"kubernetes.io/projected/ccfd4958-948b-4f5b-86ee-1c0c5242df17-kube-api-access-rtfhb\") pod \"glance-db-create-nmgcz\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.364571 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-59e4-account-create-update-wfrc2"] Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.459931 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1ec54ce-5118-4f19-b962-a4e0100f08e6-operator-scripts\") pod \"glance-59e4-account-create-update-wfrc2\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.460028 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kvvcs\" (UniqueName: \"kubernetes.io/projected/b1ec54ce-5118-4f19-b962-a4e0100f08e6-kube-api-access-kvvcs\") pod \"glance-59e4-account-create-update-wfrc2\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.460068 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rtfhb\" (UniqueName: \"kubernetes.io/projected/ccfd4958-948b-4f5b-86ee-1c0c5242df17-kube-api-access-rtfhb\") pod \"glance-db-create-nmgcz\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.460156 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfd4958-948b-4f5b-86ee-1c0c5242df17-operator-scripts\") pod \"glance-db-create-nmgcz\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.461157 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfd4958-948b-4f5b-86ee-1c0c5242df17-operator-scripts\") pod \"glance-db-create-nmgcz\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.461177 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1ec54ce-5118-4f19-b962-a4e0100f08e6-operator-scripts\") pod \"glance-59e4-account-create-update-wfrc2\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.502956 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kvvcs\" (UniqueName: \"kubernetes.io/projected/b1ec54ce-5118-4f19-b962-a4e0100f08e6-kube-api-access-kvvcs\") pod \"glance-59e4-account-create-update-wfrc2\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.533999 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rtfhb\" (UniqueName: \"kubernetes.io/projected/ccfd4958-948b-4f5b-86ee-1c0c5242df17-kube-api-access-rtfhb\") pod \"glance-db-create-nmgcz\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.624286 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:08 crc kubenswrapper[4956]: I1126 17:11:08.654608 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.188754 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-59e4-account-create-update-wfrc2"] Nov 26 17:11:09 crc kubenswrapper[4956]: W1126 17:11:09.191467 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb1ec54ce_5118_4f19_b962_a4e0100f08e6.slice/crio-d10c057bef73649d84bc83deeee08c29de9d937e968a44a67b1b096ed17d46d0 WatchSource:0}: Error finding container d10c057bef73649d84bc83deeee08c29de9d937e968a44a67b1b096ed17d46d0: Status 404 returned error can't find the container with id d10c057bef73649d84bc83deeee08c29de9d937e968a44a67b1b096ed17d46d0 Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.226204 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-nmgcz"] Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.522295 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" event={"ID":"b1ec54ce-5118-4f19-b962-a4e0100f08e6","Type":"ContainerStarted","Data":"03efd86766c47d8d2c6d010d2cb83100c25f1eb5e24996df860ea40aad397ec6"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.522725 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" event={"ID":"b1ec54ce-5118-4f19-b962-a4e0100f08e6","Type":"ContainerStarted","Data":"d10c057bef73649d84bc83deeee08c29de9d937e968a44a67b1b096ed17d46d0"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.523555 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nmgcz" event={"ID":"ccfd4958-948b-4f5b-86ee-1c0c5242df17","Type":"ContainerStarted","Data":"697de2465d9a9a21cb91c05eeda2c427180481c27c0e0f519e2a2949e4141996"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.523575 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nmgcz" event={"ID":"ccfd4958-948b-4f5b-86ee-1c0c5242df17","Type":"ContainerStarted","Data":"a39b6f43adac56a794f072e5946eb4e2e03a55355ff724ca2827be63cd285078"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.537248 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"ad37656dcf3920488183b22e10674d1c0bb386b64a716c767a7dd9a8ce748274"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.537312 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"696e92ad255ceadde77eb0dfbe4c54a8e93453ea7a58e3664eefb82e8c9b8123"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.537328 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"34ad505e66bdf6f8c7dc871b0a38ae0c67ed83bb0f6c8880193b98d2746d54d0"} Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.550925 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.551013 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.551733 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" podStartSLOduration=1.551716595 podStartE2EDuration="1.551716595s" podCreationTimestamp="2025-11-26 17:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:11:09.546108065 +0000 UTC m=+1115.242068627" watchObservedRunningTime="2025-11-26 17:11:09.551716595 +0000 UTC m=+1115.247677157" Nov 26 17:11:09 crc kubenswrapper[4956]: I1126 17:11:09.576062 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-nmgcz" podStartSLOduration=1.5760348400000002 podStartE2EDuration="1.57603484s" podCreationTimestamp="2025-11-26 17:11:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:11:09.571289124 +0000 UTC m=+1115.267249686" watchObservedRunningTime="2025-11-26 17:11:09.57603484 +0000 UTC m=+1115.271995392" Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.548817 4956 generic.go:334] "Generic (PLEG): container finished" podID="b1ec54ce-5118-4f19-b962-a4e0100f08e6" containerID="03efd86766c47d8d2c6d010d2cb83100c25f1eb5e24996df860ea40aad397ec6" exitCode=0 Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.549224 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" event={"ID":"b1ec54ce-5118-4f19-b962-a4e0100f08e6","Type":"ContainerDied","Data":"03efd86766c47d8d2c6d010d2cb83100c25f1eb5e24996df860ea40aad397ec6"} Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.551427 4956 generic.go:334] "Generic (PLEG): container finished" podID="ccfd4958-948b-4f5b-86ee-1c0c5242df17" containerID="697de2465d9a9a21cb91c05eeda2c427180481c27c0e0f519e2a2949e4141996" exitCode=0 Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.551473 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nmgcz" event={"ID":"ccfd4958-948b-4f5b-86ee-1c0c5242df17","Type":"ContainerDied","Data":"697de2465d9a9a21cb91c05eeda2c427180481c27c0e0f519e2a2949e4141996"} Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.559075 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"8afc86e2fe56c501c2ff88c1bbc040286564b6be246ac07f1b81b42a7fcf9e60"} Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.559132 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"edde7ee3972a8bba8a92e4eced4447f6dddf7df20692ff8f3c2051f35db15670"} Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.559149 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"101c803b017aa05889df18e740368e719af8483be98e964c84cbab9d5833fabd"} Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.559164 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/swift-storage-0" event={"ID":"26554e9d-3393-425b-bc71-07e66397ca3b","Type":"ContainerStarted","Data":"ade84937fa8f700c0c08eb65e0a51c186de653bda0266b9b71e879c3175f8ffa"} Nov 26 17:11:10 crc kubenswrapper[4956]: I1126 17:11:10.609657 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/swift-storage-0" podStartSLOduration=35.443703581 podStartE2EDuration="47.609634247s" podCreationTimestamp="2025-11-26 17:10:23 +0000 UTC" firstStartedPulling="2025-11-26 17:10:56.774941929 +0000 UTC m=+1102.470902491" lastFinishedPulling="2025-11-26 17:11:08.940872605 +0000 UTC m=+1114.636833157" observedRunningTime="2025-11-26 17:11:10.608827645 +0000 UTC m=+1116.304788207" watchObservedRunningTime="2025-11-26 17:11:10.609634247 +0000 UTC m=+1116.305594799" Nov 26 17:11:11 crc kubenswrapper[4956]: I1126 17:11:11.949388 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:11 crc kubenswrapper[4956]: I1126 17:11:11.958608 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.128598 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfd4958-948b-4f5b-86ee-1c0c5242df17-operator-scripts\") pod \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.129202 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1ec54ce-5118-4f19-b962-a4e0100f08e6-operator-scripts\") pod \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.129202 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ccfd4958-948b-4f5b-86ee-1c0c5242df17-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ccfd4958-948b-4f5b-86ee-1c0c5242df17" (UID: "ccfd4958-948b-4f5b-86ee-1c0c5242df17"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.129309 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kvvcs\" (UniqueName: \"kubernetes.io/projected/b1ec54ce-5118-4f19-b962-a4e0100f08e6-kube-api-access-kvvcs\") pod \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\" (UID: \"b1ec54ce-5118-4f19-b962-a4e0100f08e6\") " Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.129412 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rtfhb\" (UniqueName: \"kubernetes.io/projected/ccfd4958-948b-4f5b-86ee-1c0c5242df17-kube-api-access-rtfhb\") pod \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\" (UID: \"ccfd4958-948b-4f5b-86ee-1c0c5242df17\") " Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.129792 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b1ec54ce-5118-4f19-b962-a4e0100f08e6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "b1ec54ce-5118-4f19-b962-a4e0100f08e6" (UID: "b1ec54ce-5118-4f19-b962-a4e0100f08e6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.129984 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ccfd4958-948b-4f5b-86ee-1c0c5242df17-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.136408 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b1ec54ce-5118-4f19-b962-a4e0100f08e6-kube-api-access-kvvcs" (OuterVolumeSpecName: "kube-api-access-kvvcs") pod "b1ec54ce-5118-4f19-b962-a4e0100f08e6" (UID: "b1ec54ce-5118-4f19-b962-a4e0100f08e6"). InnerVolumeSpecName "kube-api-access-kvvcs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.141175 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ccfd4958-948b-4f5b-86ee-1c0c5242df17-kube-api-access-rtfhb" (OuterVolumeSpecName: "kube-api-access-rtfhb") pod "ccfd4958-948b-4f5b-86ee-1c0c5242df17" (UID: "ccfd4958-948b-4f5b-86ee-1c0c5242df17"). InnerVolumeSpecName "kube-api-access-rtfhb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.231215 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/b1ec54ce-5118-4f19-b962-a4e0100f08e6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.231255 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kvvcs\" (UniqueName: \"kubernetes.io/projected/b1ec54ce-5118-4f19-b962-a4e0100f08e6-kube-api-access-kvvcs\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.231269 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rtfhb\" (UniqueName: \"kubernetes.io/projected/ccfd4958-948b-4f5b-86ee-1c0c5242df17-kube-api-access-rtfhb\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.577153 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-nmgcz" event={"ID":"ccfd4958-948b-4f5b-86ee-1c0c5242df17","Type":"ContainerDied","Data":"a39b6f43adac56a794f072e5946eb4e2e03a55355ff724ca2827be63cd285078"} Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.577195 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-nmgcz" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.577212 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a39b6f43adac56a794f072e5946eb4e2e03a55355ff724ca2827be63cd285078" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.578472 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" event={"ID":"b1ec54ce-5118-4f19-b962-a4e0100f08e6","Type":"ContainerDied","Data":"d10c057bef73649d84bc83deeee08c29de9d937e968a44a67b1b096ed17d46d0"} Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.578496 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d10c057bef73649d84bc83deeee08c29de9d937e968a44a67b1b096ed17d46d0" Nov 26 17:11:12 crc kubenswrapper[4956]: I1126 17:11:12.578568 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-59e4-account-create-update-wfrc2" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.496707 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-74pwc"] Nov 26 17:11:13 crc kubenswrapper[4956]: E1126 17:11:13.497153 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ccfd4958-948b-4f5b-86ee-1c0c5242df17" containerName="mariadb-database-create" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.497169 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ccfd4958-948b-4f5b-86ee-1c0c5242df17" containerName="mariadb-database-create" Nov 26 17:11:13 crc kubenswrapper[4956]: E1126 17:11:13.497201 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b1ec54ce-5118-4f19-b962-a4e0100f08e6" containerName="mariadb-account-create-update" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.497209 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b1ec54ce-5118-4f19-b962-a4e0100f08e6" containerName="mariadb-account-create-update" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.497383 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b1ec54ce-5118-4f19-b962-a4e0100f08e6" containerName="mariadb-account-create-update" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.497405 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ccfd4958-948b-4f5b-86ee-1c0c5242df17" containerName="mariadb-database-create" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.498057 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.500962 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.501861 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qhfdm" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.504194 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.512203 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-74pwc"] Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.659139 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-db-sync-config-data\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.659220 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-combined-ca-bundle\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.659323 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-config-data\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.659733 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg54f\" (UniqueName: \"kubernetes.io/projected/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-kube-api-access-qg54f\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.760563 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg54f\" (UniqueName: \"kubernetes.io/projected/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-kube-api-access-qg54f\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.760679 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-db-sync-config-data\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.760700 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-combined-ca-bundle\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.760731 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-config-data\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.767939 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-config-data\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.771135 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-combined-ca-bundle\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.775613 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-db-sync-config-data\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.777441 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg54f\" (UniqueName: \"kubernetes.io/projected/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-kube-api-access-qg54f\") pod \"glance-db-sync-74pwc\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:13 crc kubenswrapper[4956]: I1126 17:11:13.814391 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:14 crc kubenswrapper[4956]: I1126 17:11:14.262856 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-74pwc"] Nov 26 17:11:14 crc kubenswrapper[4956]: I1126 17:11:14.596098 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-74pwc" event={"ID":"4026a0ab-d6f3-4f43-8884-7ecd3c32b522","Type":"ContainerStarted","Data":"8cc9c78ac787439f2da469c57976c145147f8b94dff73bbc7a1b13dcb4102a29"} Nov 26 17:11:26 crc kubenswrapper[4956]: I1126 17:11:26.739359 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-74pwc" event={"ID":"4026a0ab-d6f3-4f43-8884-7ecd3c32b522","Type":"ContainerStarted","Data":"33989093ae93333ad65ceb20f287cc009cdef1a1dfd04c6469371e1745fec63c"} Nov 26 17:11:26 crc kubenswrapper[4956]: I1126 17:11:26.763069 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-74pwc" podStartSLOduration=2.5285079059999997 podStartE2EDuration="13.763048477s" podCreationTimestamp="2025-11-26 17:11:13 +0000 UTC" firstStartedPulling="2025-11-26 17:11:14.271176088 +0000 UTC m=+1119.967136650" lastFinishedPulling="2025-11-26 17:11:25.505716669 +0000 UTC m=+1131.201677221" observedRunningTime="2025-11-26 17:11:26.757829898 +0000 UTC m=+1132.453790500" watchObservedRunningTime="2025-11-26 17:11:26.763048477 +0000 UTC m=+1132.459009029" Nov 26 17:11:34 crc kubenswrapper[4956]: I1126 17:11:34.799027 4956 generic.go:334] "Generic (PLEG): container finished" podID="4026a0ab-d6f3-4f43-8884-7ecd3c32b522" containerID="33989093ae93333ad65ceb20f287cc009cdef1a1dfd04c6469371e1745fec63c" exitCode=0 Nov 26 17:11:34 crc kubenswrapper[4956]: I1126 17:11:34.799089 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-74pwc" event={"ID":"4026a0ab-d6f3-4f43-8884-7ecd3c32b522","Type":"ContainerDied","Data":"33989093ae93333ad65ceb20f287cc009cdef1a1dfd04c6469371e1745fec63c"} Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.148966 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.288650 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg54f\" (UniqueName: \"kubernetes.io/projected/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-kube-api-access-qg54f\") pod \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.288824 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-db-sync-config-data\") pod \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.288921 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-config-data\") pod \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.289007 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-combined-ca-bundle\") pod \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\" (UID: \"4026a0ab-d6f3-4f43-8884-7ecd3c32b522\") " Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.296660 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-kube-api-access-qg54f" (OuterVolumeSpecName: "kube-api-access-qg54f") pod "4026a0ab-d6f3-4f43-8884-7ecd3c32b522" (UID: "4026a0ab-d6f3-4f43-8884-7ecd3c32b522"). InnerVolumeSpecName "kube-api-access-qg54f". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.300988 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "4026a0ab-d6f3-4f43-8884-7ecd3c32b522" (UID: "4026a0ab-d6f3-4f43-8884-7ecd3c32b522"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.325747 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4026a0ab-d6f3-4f43-8884-7ecd3c32b522" (UID: "4026a0ab-d6f3-4f43-8884-7ecd3c32b522"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.349040 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-config-data" (OuterVolumeSpecName: "config-data") pod "4026a0ab-d6f3-4f43-8884-7ecd3c32b522" (UID: "4026a0ab-d6f3-4f43-8884-7ecd3c32b522"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.390652 4956 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.390695 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg54f\" (UniqueName: \"kubernetes.io/projected/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-kube-api-access-qg54f\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.390710 4956 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.390721 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4026a0ab-d6f3-4f43-8884-7ecd3c32b522-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.818144 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-74pwc" event={"ID":"4026a0ab-d6f3-4f43-8884-7ecd3c32b522","Type":"ContainerDied","Data":"8cc9c78ac787439f2da469c57976c145147f8b94dff73bbc7a1b13dcb4102a29"} Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.818197 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8cc9c78ac787439f2da469c57976c145147f8b94dff73bbc7a1b13dcb4102a29" Nov 26 17:11:36 crc kubenswrapper[4956]: I1126 17:11:36.818287 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-74pwc" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.279027 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:38 crc kubenswrapper[4956]: E1126 17:11:38.279778 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4026a0ab-d6f3-4f43-8884-7ecd3c32b522" containerName="glance-db-sync" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.279792 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4026a0ab-d6f3-4f43-8884-7ecd3c32b522" containerName="glance-db-sync" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.280034 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4026a0ab-d6f3-4f43-8884-7ecd3c32b522" containerName="glance-db-sync" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.280973 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.285733 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.286347 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qhfdm" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.286578 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.286762 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.286903 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.286909 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.294324 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.425931 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-scripts\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.425982 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426012 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426041 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g8vs5\" (UniqueName: \"kubernetes.io/projected/c039adce-2d7a-47c2-90cf-273928886ee3-kube-api-access-g8vs5\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426130 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426241 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-config-data\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426354 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426464 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-logs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.426560 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-httpd-run\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528155 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528221 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-scripts\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528258 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528289 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g8vs5\" (UniqueName: \"kubernetes.io/projected/c039adce-2d7a-47c2-90cf-273928886ee3-kube-api-access-g8vs5\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528341 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528368 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-config-data\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528402 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528450 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-logs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.528500 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-httpd-run\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.529209 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-httpd-run\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.529416 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.529728 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-logs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.535251 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-scripts\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.536848 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.536897 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.542134 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-config-data\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.543758 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.547320 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g8vs5\" (UniqueName: \"kubernetes.io/projected/c039adce-2d7a-47c2-90cf-273928886ee3-kube-api-access-g8vs5\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.550935 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:38 crc kubenswrapper[4956]: I1126 17:11:38.602392 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:39 crc kubenswrapper[4956]: I1126 17:11:39.551030 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:11:39 crc kubenswrapper[4956]: I1126 17:11:39.551114 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:11:43 crc kubenswrapper[4956]: I1126 17:11:43.877176 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:51 crc kubenswrapper[4956]: I1126 17:11:51.742296 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:51 crc kubenswrapper[4956]: I1126 17:11:51.958903 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c039adce-2d7a-47c2-90cf-273928886ee3","Type":"ContainerStarted","Data":"52e8ed42c4d171d64c5ba28c35bf66ecb32d875b39e9bd2f3b4b678015caffc1"} Nov 26 17:11:52 crc kubenswrapper[4956]: I1126 17:11:52.970214 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c039adce-2d7a-47c2-90cf-273928886ee3","Type":"ContainerStarted","Data":"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b"} Nov 26 17:11:52 crc kubenswrapper[4956]: I1126 17:11:52.970411 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-log" containerID="cri-o://c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b" gracePeriod=30 Nov 26 17:11:52 crc kubenswrapper[4956]: I1126 17:11:52.970454 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-httpd" containerID="cri-o://d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b" gracePeriod=30 Nov 26 17:11:52 crc kubenswrapper[4956]: I1126 17:11:52.972102 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c039adce-2d7a-47c2-90cf-273928886ee3","Type":"ContainerStarted","Data":"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b"} Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.017138 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=16.017111995 podStartE2EDuration="16.017111995s" podCreationTimestamp="2025-11-26 17:11:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:11:52.999527103 +0000 UTC m=+1158.695487655" watchObservedRunningTime="2025-11-26 17:11:53.017111995 +0000 UTC m=+1158.713072547" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.321627 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.479845 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-public-tls-certs\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.479958 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-scripts\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480007 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-config-data\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480030 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480051 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g8vs5\" (UniqueName: \"kubernetes.io/projected/c039adce-2d7a-47c2-90cf-273928886ee3-kube-api-access-g8vs5\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480101 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-combined-ca-bundle\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480177 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-internal-tls-certs\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480249 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-httpd-run\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.480277 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-logs\") pod \"c039adce-2d7a-47c2-90cf-273928886ee3\" (UID: \"c039adce-2d7a-47c2-90cf-273928886ee3\") " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.481166 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-logs" (OuterVolumeSpecName: "logs") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.481180 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.487232 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-scripts" (OuterVolumeSpecName: "scripts") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.487362 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.488154 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c039adce-2d7a-47c2-90cf-273928886ee3-kube-api-access-g8vs5" (OuterVolumeSpecName: "kube-api-access-g8vs5") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "kube-api-access-g8vs5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.506707 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.519129 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.522919 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.530286 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-config-data" (OuterVolumeSpecName: "config-data") pod "c039adce-2d7a-47c2-90cf-273928886ee3" (UID: "c039adce-2d7a-47c2-90cf-273928886ee3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581675 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581749 4956 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581766 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581780 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581827 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581845 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g8vs5\" (UniqueName: \"kubernetes.io/projected/c039adce-2d7a-47c2-90cf-273928886ee3-kube-api-access-g8vs5\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581874 4956 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581890 4956 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c039adce-2d7a-47c2-90cf-273928886ee3-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.581903 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/c039adce-2d7a-47c2-90cf-273928886ee3-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.596384 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.683382 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980419 4956 generic.go:334] "Generic (PLEG): container finished" podID="c039adce-2d7a-47c2-90cf-273928886ee3" containerID="d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b" exitCode=143 Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980461 4956 generic.go:334] "Generic (PLEG): container finished" podID="c039adce-2d7a-47c2-90cf-273928886ee3" containerID="c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b" exitCode=143 Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980488 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c039adce-2d7a-47c2-90cf-273928886ee3","Type":"ContainerDied","Data":"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b"} Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980499 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980522 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c039adce-2d7a-47c2-90cf-273928886ee3","Type":"ContainerDied","Data":"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b"} Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980536 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"c039adce-2d7a-47c2-90cf-273928886ee3","Type":"ContainerDied","Data":"52e8ed42c4d171d64c5ba28c35bf66ecb32d875b39e9bd2f3b4b678015caffc1"} Nov 26 17:11:53 crc kubenswrapper[4956]: I1126 17:11:53.980555 4956 scope.go:117] "RemoveContainer" containerID="d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.015068 4956 scope.go:117] "RemoveContainer" containerID="c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.018329 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.025722 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.037437 4956 scope.go:117] "RemoveContainer" containerID="d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b" Nov 26 17:11:54 crc kubenswrapper[4956]: E1126 17:11:54.043068 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b\": container with ID starting with d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b not found: ID does not exist" containerID="d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.043123 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b"} err="failed to get container status \"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b\": rpc error: code = NotFound desc = could not find container \"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b\": container with ID starting with d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b not found: ID does not exist" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.043163 4956 scope.go:117] "RemoveContainer" containerID="c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b" Nov 26 17:11:54 crc kubenswrapper[4956]: E1126 17:11:54.043851 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b\": container with ID starting with c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b not found: ID does not exist" containerID="c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.044260 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b"} err="failed to get container status \"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b\": rpc error: code = NotFound desc = could not find container \"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b\": container with ID starting with c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b not found: ID does not exist" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.044297 4956 scope.go:117] "RemoveContainer" containerID="d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.044818 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b"} err="failed to get container status \"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b\": rpc error: code = NotFound desc = could not find container \"d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b\": container with ID starting with d8605803b7581d00b9fd8bd5480f2a196b558f2bc0d2dd7f32f74f8bc69fe66b not found: ID does not exist" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.044847 4956 scope.go:117] "RemoveContainer" containerID="c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.045612 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b"} err="failed to get container status \"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b\": rpc error: code = NotFound desc = could not find container \"c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b\": container with ID starting with c8676574c1c882eccd5c8b90fba406bff25e3ae2fbea723bc65843d18a71585b not found: ID does not exist" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.058145 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:54 crc kubenswrapper[4956]: E1126 17:11:54.058599 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-httpd" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.058617 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-httpd" Nov 26 17:11:54 crc kubenswrapper[4956]: E1126 17:11:54.058630 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-log" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.058637 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-log" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.058783 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-httpd" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.058797 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" containerName="glance-log" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.059800 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.062597 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-public-svc" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.063071 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.064058 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.064422 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qhfdm" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.064949 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.065065 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"cert-glance-default-internal-svc" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.065141 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"combined-ca-bundle" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195101 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195169 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195224 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-httpd-run\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195277 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195368 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b79sn\" (UniqueName: \"kubernetes.io/projected/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-kube-api-access-b79sn\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195438 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195469 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195493 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-logs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.195517 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.297124 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-httpd-run\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.297808 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-httpd-run\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.297219 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.297941 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b79sn\" (UniqueName: \"kubernetes.io/projected/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-kube-api-access-b79sn\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.298401 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.298921 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.298967 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-logs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.299002 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.299124 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.299274 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.299494 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-logs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.299849 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.304092 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-combined-ca-bundle\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.304211 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-internal-tls-certs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.304341 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.305005 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-public-tls-certs\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.307996 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.316661 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b79sn\" (UniqueName: \"kubernetes.io/projected/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-kube-api-access-b79sn\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.332950 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-0\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.392168 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:11:54 crc kubenswrapper[4956]: I1126 17:11:54.859232 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:11:54 crc kubenswrapper[4956]: W1126 17:11:54.867350 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod27cef2b9_104b_42b4_9ac4_fdb10a6c0473.slice/crio-e2a9b81fffe1edb51f7dcd7eb07de4b02a443913becc5f6e329207680751546d WatchSource:0}: Error finding container e2a9b81fffe1edb51f7dcd7eb07de4b02a443913becc5f6e329207680751546d: Status 404 returned error can't find the container with id e2a9b81fffe1edb51f7dcd7eb07de4b02a443913becc5f6e329207680751546d Nov 26 17:11:55 crc kubenswrapper[4956]: I1126 17:11:55.011093 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c039adce-2d7a-47c2-90cf-273928886ee3" path="/var/lib/kubelet/pods/c039adce-2d7a-47c2-90cf-273928886ee3/volumes" Nov 26 17:11:55 crc kubenswrapper[4956]: I1126 17:11:55.012040 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"27cef2b9-104b-42b4-9ac4-fdb10a6c0473","Type":"ContainerStarted","Data":"e2a9b81fffe1edb51f7dcd7eb07de4b02a443913becc5f6e329207680751546d"} Nov 26 17:11:56 crc kubenswrapper[4956]: I1126 17:11:56.008462 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"27cef2b9-104b-42b4-9ac4-fdb10a6c0473","Type":"ContainerStarted","Data":"8ce45d87139417a39d3868e85480ad79841033a308b69b276463a679af2f3994"} Nov 26 17:11:56 crc kubenswrapper[4956]: I1126 17:11:56.008886 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"27cef2b9-104b-42b4-9ac4-fdb10a6c0473","Type":"ContainerStarted","Data":"452736cef2c704083546d22717869be3fa89234d6c5b1cdb1884b8b59c582b2f"} Nov 26 17:12:04 crc kubenswrapper[4956]: I1126 17:12:04.392575 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:04 crc kubenswrapper[4956]: I1126 17:12:04.393713 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:04 crc kubenswrapper[4956]: I1126 17:12:04.441312 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:04 crc kubenswrapper[4956]: I1126 17:12:04.449823 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:04 crc kubenswrapper[4956]: I1126 17:12:04.487118 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=10.487090172 podStartE2EDuration="10.487090172s" podCreationTimestamp="2025-11-26 17:11:54 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:11:56.044218671 +0000 UTC m=+1161.740179223" watchObservedRunningTime="2025-11-26 17:12:04.487090172 +0000 UTC m=+1170.183050754" Nov 26 17:12:05 crc kubenswrapper[4956]: I1126 17:12:05.093246 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:05 crc kubenswrapper[4956]: I1126 17:12:05.093785 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:07 crc kubenswrapper[4956]: I1126 17:12:07.299720 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:07 crc kubenswrapper[4956]: I1126 17:12:07.299902 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:12:07 crc kubenswrapper[4956]: I1126 17:12:07.301905 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.201117 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-74pwc"] Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.206740 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-74pwc"] Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.257419 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance59e4-account-delete-dflng"] Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.258566 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.262651 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.262751 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:08.762726172 +0000 UTC m=+1174.458686724 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-default-single-config-data" not found Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.263233 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.263337 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:08.763308479 +0000 UTC m=+1174.459269021 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-scripts" not found Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.274210 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance59e4-account-delete-dflng"] Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.362915 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-operator-scripts\") pod \"glance59e4-account-delete-dflng\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.363195 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92tls\" (UniqueName: \"kubernetes.io/projected/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-kube-api-access-92tls\") pod \"glance59e4-account-delete-dflng\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.377745 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.465522 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-operator-scripts\") pod \"glance59e4-account-delete-dflng\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.465628 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92tls\" (UniqueName: \"kubernetes.io/projected/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-kube-api-access-92tls\") pod \"glance59e4-account-delete-dflng\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.466763 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-operator-scripts\") pod \"glance59e4-account-delete-dflng\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.486756 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92tls\" (UniqueName: \"kubernetes.io/projected/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-kube-api-access-92tls\") pod \"glance59e4-account-delete-dflng\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.646368 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.774039 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.774446 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:09.77442361 +0000 UTC m=+1175.470384162 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-default-single-config-data" not found Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.774060 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 26 17:12:08 crc kubenswrapper[4956]: E1126 17:12:08.775572 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:09.775561893 +0000 UTC m=+1175.471522445 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-scripts" not found Nov 26 17:12:08 crc kubenswrapper[4956]: I1126 17:12:08.918522 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance59e4-account-delete-dflng"] Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.061967 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4026a0ab-d6f3-4f43-8884-7ecd3c32b522" path="/var/lib/kubelet/pods/4026a0ab-d6f3-4f43-8884-7ecd3c32b522/volumes" Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.137054 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-log" containerID="cri-o://452736cef2c704083546d22717869be3fa89234d6c5b1cdb1884b8b59c582b2f" gracePeriod=30 Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.137150 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" event={"ID":"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89","Type":"ContainerStarted","Data":"39dcf839f7906cb527e52239a876f0db2a3a1b1f80d0e996994a55a5324e0b51"} Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.137539 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-httpd" containerID="cri-o://8ce45d87139417a39d3868e85480ad79841033a308b69b276463a679af2f3994" gracePeriod=30 Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.164193 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-log" probeResult="failure" output="Get \"https://10.217.0.99:9292/healthcheck\": EOF" Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.551781 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.552344 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.552445 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.553552 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e26914562ac9efcf004522ee5db12eb4d489e28d6cbf474316c25c32bd557c51"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 17:12:09 crc kubenswrapper[4956]: I1126 17:12:09.553615 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://e26914562ac9efcf004522ee5db12eb4d489e28d6cbf474316c25c32bd557c51" gracePeriod=600 Nov 26 17:12:09 crc kubenswrapper[4956]: E1126 17:12:09.800384 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 26 17:12:09 crc kubenswrapper[4956]: E1126 17:12:09.800488 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:11.800465791 +0000 UTC m=+1177.496426413 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-default-single-config-data" not found Nov 26 17:12:09 crc kubenswrapper[4956]: E1126 17:12:09.800401 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 26 17:12:09 crc kubenswrapper[4956]: E1126 17:12:09.800602 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:11.800581404 +0000 UTC m=+1177.496542026 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-scripts" not found Nov 26 17:12:10 crc kubenswrapper[4956]: I1126 17:12:10.146325 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" event={"ID":"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89","Type":"ContainerStarted","Data":"5e847741d07bb4d6f3b750c5a1b08df36f3716a2e8427ce9a73f7a0855c859a0"} Nov 26 17:12:10 crc kubenswrapper[4956]: I1126 17:12:10.148545 4956 generic.go:334] "Generic (PLEG): container finished" podID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerID="452736cef2c704083546d22717869be3fa89234d6c5b1cdb1884b8b59c582b2f" exitCode=143 Nov 26 17:12:10 crc kubenswrapper[4956]: I1126 17:12:10.148584 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"27cef2b9-104b-42b4-9ac4-fdb10a6c0473","Type":"ContainerDied","Data":"452736cef2c704083546d22717869be3fa89234d6c5b1cdb1884b8b59c582b2f"} Nov 26 17:12:10 crc kubenswrapper[4956]: I1126 17:12:10.163797 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" podStartSLOduration=2.16377066 podStartE2EDuration="2.16377066s" podCreationTimestamp="2025-11-26 17:12:08 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:10.160403464 +0000 UTC m=+1175.856364026" watchObservedRunningTime="2025-11-26 17:12:10.16377066 +0000 UTC m=+1175.859731212" Nov 26 17:12:11 crc kubenswrapper[4956]: I1126 17:12:11.162019 4956 generic.go:334] "Generic (PLEG): container finished" podID="5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" containerID="5e847741d07bb4d6f3b750c5a1b08df36f3716a2e8427ce9a73f7a0855c859a0" exitCode=0 Nov 26 17:12:11 crc kubenswrapper[4956]: I1126 17:12:11.162933 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" event={"ID":"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89","Type":"ContainerDied","Data":"5e847741d07bb4d6f3b750c5a1b08df36f3716a2e8427ce9a73f7a0855c859a0"} Nov 26 17:12:11 crc kubenswrapper[4956]: I1126 17:12:11.166419 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="e26914562ac9efcf004522ee5db12eb4d489e28d6cbf474316c25c32bd557c51" exitCode=0 Nov 26 17:12:11 crc kubenswrapper[4956]: I1126 17:12:11.167183 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"e26914562ac9efcf004522ee5db12eb4d489e28d6cbf474316c25c32bd557c51"} Nov 26 17:12:11 crc kubenswrapper[4956]: I1126 17:12:11.167284 4956 scope.go:117] "RemoveContainer" containerID="0683eafa4a97fe7cebdb5b35d5b4a8e27469c373b19946cd7cb9d38e21f774da" Nov 26 17:12:11 crc kubenswrapper[4956]: E1126 17:12:11.835853 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-scripts: secret "glance-scripts" not found Nov 26 17:12:11 crc kubenswrapper[4956]: E1126 17:12:11.836197 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:15.836180757 +0000 UTC m=+1181.532141299 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "scripts" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-scripts" not found Nov 26 17:12:11 crc kubenswrapper[4956]: E1126 17:12:11.835941 4956 secret.go:188] Couldn't get secret glance-kuttl-tests/glance-default-single-config-data: secret "glance-default-single-config-data" not found Nov 26 17:12:11 crc kubenswrapper[4956]: E1126 17:12:11.836342 4956 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data podName:27cef2b9-104b-42b4-9ac4-fdb10a6c0473 nodeName:}" failed. No retries permitted until 2025-11-26 17:12:15.8363111 +0000 UTC m=+1181.532271672 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "config-data" (UniqueName: "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data") pod "glance-default-single-0" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473") : secret "glance-default-single-config-data" not found Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.178518 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"6afa66b60ccff62ba0cee9768e910db0d5a65279fb098588258f5b5dc187f77e"} Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.515495 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.654045 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-operator-scripts\") pod \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.654265 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92tls\" (UniqueName: \"kubernetes.io/projected/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-kube-api-access-92tls\") pod \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\" (UID: \"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89\") " Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.656321 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" (UID: "5ad7619a-cb73-4c01-ac89-f98dd6ff2f89"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.664605 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-kube-api-access-92tls" (OuterVolumeSpecName: "kube-api-access-92tls") pod "5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" (UID: "5ad7619a-cb73-4c01-ac89-f98dd6ff2f89"). InnerVolumeSpecName "kube-api-access-92tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.757461 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92tls\" (UniqueName: \"kubernetes.io/projected/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-kube-api-access-92tls\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:12 crc kubenswrapper[4956]: I1126 17:12:12.757509 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.187833 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" event={"ID":"5ad7619a-cb73-4c01-ac89-f98dd6ff2f89","Type":"ContainerDied","Data":"39dcf839f7906cb527e52239a876f0db2a3a1b1f80d0e996994a55a5324e0b51"} Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.187892 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance59e4-account-delete-dflng" Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.188498 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="39dcf839f7906cb527e52239a876f0db2a3a1b1f80d0e996994a55a5324e0b51" Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.278411 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-nmgcz"] Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.287492 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-nmgcz"] Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.307912 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance59e4-account-delete-dflng"] Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.312859 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-59e4-account-create-update-wfrc2"] Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.317090 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance59e4-account-delete-dflng"] Nov 26 17:12:13 crc kubenswrapper[4956]: I1126 17:12:13.321568 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-59e4-account-create-update-wfrc2"] Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.197952 4956 generic.go:334] "Generic (PLEG): container finished" podID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerID="8ce45d87139417a39d3868e85480ad79841033a308b69b276463a679af2f3994" exitCode=0 Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.198010 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"27cef2b9-104b-42b4-9ac4-fdb10a6c0473","Type":"ContainerDied","Data":"8ce45d87139417a39d3868e85480ad79841033a308b69b276463a679af2f3994"} Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.580433 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690098 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b79sn\" (UniqueName: \"kubernetes.io/projected/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-kube-api-access-b79sn\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690163 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-combined-ca-bundle\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690200 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690235 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-logs\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690268 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690296 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-internal-tls-certs\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690368 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690402 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-httpd-run\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.690425 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-public-tls-certs\") pod \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\" (UID: \"27cef2b9-104b-42b4-9ac4-fdb10a6c0473\") " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.691817 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.691843 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-logs" (OuterVolumeSpecName: "logs") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.700731 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.701046 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts" (OuterVolumeSpecName: "scripts") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.706346 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-kube-api-access-b79sn" (OuterVolumeSpecName: "kube-api-access-b79sn") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "kube-api-access-b79sn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.725103 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.732208 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.741781 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.745515 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data" (OuterVolumeSpecName: "config-data") pod "27cef2b9-104b-42b4-9ac4-fdb10a6c0473" (UID: "27cef2b9-104b-42b4-9ac4-fdb10a6c0473"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792811 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b79sn\" (UniqueName: \"kubernetes.io/projected/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-kube-api-access-b79sn\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792885 4956 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792900 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792913 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792953 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792963 4956 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792972 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792980 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.792988 4956 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/27cef2b9-104b-42b4-9ac4-fdb10a6c0473-public-tls-certs\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.808012 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:12:14 crc kubenswrapper[4956]: I1126 17:12:14.894994 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:15 crc kubenswrapper[4956]: I1126 17:12:15.005702 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" path="/var/lib/kubelet/pods/5ad7619a-cb73-4c01-ac89-f98dd6ff2f89/volumes" Nov 26 17:12:15 crc kubenswrapper[4956]: I1126 17:12:15.006760 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b1ec54ce-5118-4f19-b962-a4e0100f08e6" path="/var/lib/kubelet/pods/b1ec54ce-5118-4f19-b962-a4e0100f08e6/volumes" Nov 26 17:12:15 crc kubenswrapper[4956]: I1126 17:12:15.008933 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ccfd4958-948b-4f5b-86ee-1c0c5242df17" path="/var/lib/kubelet/pods/ccfd4958-948b-4f5b-86ee-1c0c5242df17/volumes" Nov 26 17:12:15 crc kubenswrapper[4956]: I1126 17:12:15.210674 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"27cef2b9-104b-42b4-9ac4-fdb10a6c0473","Type":"ContainerDied","Data":"e2a9b81fffe1edb51f7dcd7eb07de4b02a443913becc5f6e329207680751546d"} Nov 26 17:12:15 crc kubenswrapper[4956]: I1126 17:12:15.210744 4956 scope.go:117] "RemoveContainer" containerID="8ce45d87139417a39d3868e85480ad79841033a308b69b276463a679af2f3994" Nov 26 17:12:15 crc kubenswrapper[4956]: I1126 17:12:15.210752 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:15.243942 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:15.250985 4956 scope.go:117] "RemoveContainer" containerID="452736cef2c704083546d22717869be3fa89234d6c5b1cdb1884b8b59c582b2f" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:15.252402 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306188 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-hhkv8"] Nov 26 17:12:16 crc kubenswrapper[4956]: E1126 17:12:16.306495 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-httpd" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306508 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-httpd" Nov 26 17:12:16 crc kubenswrapper[4956]: E1126 17:12:16.306529 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" containerName="mariadb-account-delete" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306535 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" containerName="mariadb-account-delete" Nov 26 17:12:16 crc kubenswrapper[4956]: E1126 17:12:16.306553 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-log" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306559 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-log" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306695 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-httpd" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306706 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" containerName="glance-log" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.306714 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5ad7619a-cb73-4c01-ac89-f98dd6ff2f89" containerName="mariadb-account-delete" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.307234 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.318007 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-hhkv8"] Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.422848 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1819f68-cff5-43a4-a73b-d37a47fd26e1-operator-scripts\") pod \"glance-db-create-hhkv8\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.423349 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6nbn6\" (UniqueName: \"kubernetes.io/projected/c1819f68-cff5-43a4-a73b-d37a47fd26e1-kube-api-access-6nbn6\") pod \"glance-db-create-hhkv8\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.425993 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-15f5-account-create-update-rkz5m"] Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.427049 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.430722 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.433191 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-15f5-account-create-update-rkz5m"] Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.524755 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1819f68-cff5-43a4-a73b-d37a47fd26e1-operator-scripts\") pod \"glance-db-create-hhkv8\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.524881 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjlxs\" (UniqueName: \"kubernetes.io/projected/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-kube-api-access-xjlxs\") pod \"glance-15f5-account-create-update-rkz5m\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.524926 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6nbn6\" (UniqueName: \"kubernetes.io/projected/c1819f68-cff5-43a4-a73b-d37a47fd26e1-kube-api-access-6nbn6\") pod \"glance-db-create-hhkv8\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.524966 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-operator-scripts\") pod \"glance-15f5-account-create-update-rkz5m\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.525525 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1819f68-cff5-43a4-a73b-d37a47fd26e1-operator-scripts\") pod \"glance-db-create-hhkv8\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.545777 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6nbn6\" (UniqueName: \"kubernetes.io/projected/c1819f68-cff5-43a4-a73b-d37a47fd26e1-kube-api-access-6nbn6\") pod \"glance-db-create-hhkv8\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.623475 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.626796 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjlxs\" (UniqueName: \"kubernetes.io/projected/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-kube-api-access-xjlxs\") pod \"glance-15f5-account-create-update-rkz5m\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.626938 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-operator-scripts\") pod \"glance-15f5-account-create-update-rkz5m\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.627677 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-operator-scripts\") pod \"glance-15f5-account-create-update-rkz5m\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.652957 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjlxs\" (UniqueName: \"kubernetes.io/projected/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-kube-api-access-xjlxs\") pod \"glance-15f5-account-create-update-rkz5m\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:16 crc kubenswrapper[4956]: I1126 17:12:16.758038 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:17 crc kubenswrapper[4956]: I1126 17:12:17.005068 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="27cef2b9-104b-42b4-9ac4-fdb10a6c0473" path="/var/lib/kubelet/pods/27cef2b9-104b-42b4-9ac4-fdb10a6c0473/volumes" Nov 26 17:12:17 crc kubenswrapper[4956]: I1126 17:12:17.006345 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-15f5-account-create-update-rkz5m"] Nov 26 17:12:17 crc kubenswrapper[4956]: I1126 17:12:17.083430 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-hhkv8"] Nov 26 17:12:17 crc kubenswrapper[4956]: W1126 17:12:17.089519 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1819f68_cff5_43a4_a73b_d37a47fd26e1.slice/crio-9f5956384b7c28725face8d7e7adbdcb526889e2c9983fd77efcb10ca6e9a74f WatchSource:0}: Error finding container 9f5956384b7c28725face8d7e7adbdcb526889e2c9983fd77efcb10ca6e9a74f: Status 404 returned error can't find the container with id 9f5956384b7c28725face8d7e7adbdcb526889e2c9983fd77efcb10ca6e9a74f Nov 26 17:12:17 crc kubenswrapper[4956]: I1126 17:12:17.232326 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-hhkv8" event={"ID":"c1819f68-cff5-43a4-a73b-d37a47fd26e1","Type":"ContainerStarted","Data":"9f5956384b7c28725face8d7e7adbdcb526889e2c9983fd77efcb10ca6e9a74f"} Nov 26 17:12:17 crc kubenswrapper[4956]: I1126 17:12:17.234039 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" event={"ID":"1292cc2f-50be-4bdb-b9af-f939e15ca7d6","Type":"ContainerStarted","Data":"91fdd50d39ce99fe378a855bf8b28c2f29715d7cc9e6668cd918173a59dba1bb"} Nov 26 17:12:18 crc kubenswrapper[4956]: I1126 17:12:18.246675 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-hhkv8" event={"ID":"c1819f68-cff5-43a4-a73b-d37a47fd26e1","Type":"ContainerStarted","Data":"f3475b2f7f525a60d6c12ab5fe9923fe6a3a1ef3c573347a59d8b537e5ef224c"} Nov 26 17:12:18 crc kubenswrapper[4956]: I1126 17:12:18.252081 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" event={"ID":"1292cc2f-50be-4bdb-b9af-f939e15ca7d6","Type":"ContainerStarted","Data":"038f4844efd8bfd5d6bf9e59b4e217d721c1e513b9fdf871cd8732cd20381aa1"} Nov 26 17:12:18 crc kubenswrapper[4956]: I1126 17:12:18.271128 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-hhkv8" podStartSLOduration=2.271095184 podStartE2EDuration="2.271095184s" podCreationTimestamp="2025-11-26 17:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:18.268332335 +0000 UTC m=+1183.964292927" watchObservedRunningTime="2025-11-26 17:12:18.271095184 +0000 UTC m=+1183.967055756" Nov 26 17:12:18 crc kubenswrapper[4956]: I1126 17:12:18.296652 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" podStartSLOduration=2.296621064 podStartE2EDuration="2.296621064s" podCreationTimestamp="2025-11-26 17:12:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:18.288749689 +0000 UTC m=+1183.984710251" watchObservedRunningTime="2025-11-26 17:12:18.296621064 +0000 UTC m=+1183.992581646" Nov 26 17:12:20 crc kubenswrapper[4956]: I1126 17:12:20.272476 4956 generic.go:334] "Generic (PLEG): container finished" podID="c1819f68-cff5-43a4-a73b-d37a47fd26e1" containerID="f3475b2f7f525a60d6c12ab5fe9923fe6a3a1ef3c573347a59d8b537e5ef224c" exitCode=0 Nov 26 17:12:20 crc kubenswrapper[4956]: I1126 17:12:20.272574 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-hhkv8" event={"ID":"c1819f68-cff5-43a4-a73b-d37a47fd26e1","Type":"ContainerDied","Data":"f3475b2f7f525a60d6c12ab5fe9923fe6a3a1ef3c573347a59d8b537e5ef224c"} Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.286708 4956 generic.go:334] "Generic (PLEG): container finished" podID="1292cc2f-50be-4bdb-b9af-f939e15ca7d6" containerID="038f4844efd8bfd5d6bf9e59b4e217d721c1e513b9fdf871cd8732cd20381aa1" exitCode=0 Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.286801 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" event={"ID":"1292cc2f-50be-4bdb-b9af-f939e15ca7d6","Type":"ContainerDied","Data":"038f4844efd8bfd5d6bf9e59b4e217d721c1e513b9fdf871cd8732cd20381aa1"} Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.525757 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.656174 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1819f68-cff5-43a4-a73b-d37a47fd26e1-operator-scripts\") pod \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.656401 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6nbn6\" (UniqueName: \"kubernetes.io/projected/c1819f68-cff5-43a4-a73b-d37a47fd26e1-kube-api-access-6nbn6\") pod \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\" (UID: \"c1819f68-cff5-43a4-a73b-d37a47fd26e1\") " Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.657617 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1819f68-cff5-43a4-a73b-d37a47fd26e1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1819f68-cff5-43a4-a73b-d37a47fd26e1" (UID: "c1819f68-cff5-43a4-a73b-d37a47fd26e1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.663255 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1819f68-cff5-43a4-a73b-d37a47fd26e1-kube-api-access-6nbn6" (OuterVolumeSpecName: "kube-api-access-6nbn6") pod "c1819f68-cff5-43a4-a73b-d37a47fd26e1" (UID: "c1819f68-cff5-43a4-a73b-d37a47fd26e1"). InnerVolumeSpecName "kube-api-access-6nbn6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.759077 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1819f68-cff5-43a4-a73b-d37a47fd26e1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:21 crc kubenswrapper[4956]: I1126 17:12:21.759139 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6nbn6\" (UniqueName: \"kubernetes.io/projected/c1819f68-cff5-43a4-a73b-d37a47fd26e1-kube-api-access-6nbn6\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.299858 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-hhkv8" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.301439 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-hhkv8" event={"ID":"c1819f68-cff5-43a4-a73b-d37a47fd26e1","Type":"ContainerDied","Data":"9f5956384b7c28725face8d7e7adbdcb526889e2c9983fd77efcb10ca6e9a74f"} Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.301519 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f5956384b7c28725face8d7e7adbdcb526889e2c9983fd77efcb10ca6e9a74f" Nov 26 17:12:22 crc kubenswrapper[4956]: E1126 17:12:22.421604 4956 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc1819f68_cff5_43a4_a73b_d37a47fd26e1.slice\": RecentStats: unable to find data in memory cache]" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.638192 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.674333 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-operator-scripts\") pod \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.674597 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjlxs\" (UniqueName: \"kubernetes.io/projected/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-kube-api-access-xjlxs\") pod \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\" (UID: \"1292cc2f-50be-4bdb-b9af-f939e15ca7d6\") " Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.676277 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1292cc2f-50be-4bdb-b9af-f939e15ca7d6" (UID: "1292cc2f-50be-4bdb-b9af-f939e15ca7d6"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.680836 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-kube-api-access-xjlxs" (OuterVolumeSpecName: "kube-api-access-xjlxs") pod "1292cc2f-50be-4bdb-b9af-f939e15ca7d6" (UID: "1292cc2f-50be-4bdb-b9af-f939e15ca7d6"). InnerVolumeSpecName "kube-api-access-xjlxs". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.776802 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjlxs\" (UniqueName: \"kubernetes.io/projected/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-kube-api-access-xjlxs\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:22 crc kubenswrapper[4956]: I1126 17:12:22.776887 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1292cc2f-50be-4bdb-b9af-f939e15ca7d6-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:23 crc kubenswrapper[4956]: I1126 17:12:23.310990 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" event={"ID":"1292cc2f-50be-4bdb-b9af-f939e15ca7d6","Type":"ContainerDied","Data":"91fdd50d39ce99fe378a855bf8b28c2f29715d7cc9e6668cd918173a59dba1bb"} Nov 26 17:12:23 crc kubenswrapper[4956]: I1126 17:12:23.311986 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="91fdd50d39ce99fe378a855bf8b28c2f29715d7cc9e6668cd918173a59dba1bb" Nov 26 17:12:23 crc kubenswrapper[4956]: I1126 17:12:23.311133 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-15f5-account-create-update-rkz5m" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.745316 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-pdnsb"] Nov 26 17:12:26 crc kubenswrapper[4956]: E1126 17:12:26.746774 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1819f68-cff5-43a4-a73b-d37a47fd26e1" containerName="mariadb-database-create" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.746909 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1819f68-cff5-43a4-a73b-d37a47fd26e1" containerName="mariadb-database-create" Nov 26 17:12:26 crc kubenswrapper[4956]: E1126 17:12:26.746995 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1292cc2f-50be-4bdb-b9af-f939e15ca7d6" containerName="mariadb-account-create-update" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.747054 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1292cc2f-50be-4bdb-b9af-f939e15ca7d6" containerName="mariadb-account-create-update" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.747239 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1819f68-cff5-43a4-a73b-d37a47fd26e1" containerName="mariadb-database-create" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.747318 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1292cc2f-50be-4bdb-b9af-f939e15ca7d6" containerName="mariadb-account-create-update" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.748163 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.750620 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.751344 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-vb95k" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.758035 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pdnsb"] Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.763846 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-db-sync-config-data\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.764013 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-config-data\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.764161 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9bst\" (UniqueName: \"kubernetes.io/projected/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-kube-api-access-v9bst\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.865795 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-db-sync-config-data\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.865854 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-config-data\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.865913 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9bst\" (UniqueName: \"kubernetes.io/projected/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-kube-api-access-v9bst\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.874626 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-config-data\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.884630 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-db-sync-config-data\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:26 crc kubenswrapper[4956]: I1126 17:12:26.886802 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9bst\" (UniqueName: \"kubernetes.io/projected/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-kube-api-access-v9bst\") pod \"glance-db-sync-pdnsb\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:27 crc kubenswrapper[4956]: I1126 17:12:27.066013 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:27 crc kubenswrapper[4956]: I1126 17:12:27.329830 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pdnsb"] Nov 26 17:12:27 crc kubenswrapper[4956]: I1126 17:12:27.347355 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pdnsb" event={"ID":"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c","Type":"ContainerStarted","Data":"ac26029a37d902c1345fc8947639ae285d3a38868a6ecdfddd17c339192010b4"} Nov 26 17:12:28 crc kubenswrapper[4956]: I1126 17:12:28.356224 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pdnsb" event={"ID":"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c","Type":"ContainerStarted","Data":"bb90bf30ca9aa260b87126883ac63132095877dca04b01f21bb7ea444302ab55"} Nov 26 17:12:28 crc kubenswrapper[4956]: I1126 17:12:28.379210 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-pdnsb" podStartSLOduration=2.3791937450000002 podStartE2EDuration="2.379193745s" podCreationTimestamp="2025-11-26 17:12:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:28.374673406 +0000 UTC m=+1194.070633958" watchObservedRunningTime="2025-11-26 17:12:28.379193745 +0000 UTC m=+1194.075154297" Nov 26 17:12:31 crc kubenswrapper[4956]: I1126 17:12:31.386359 4956 generic.go:334] "Generic (PLEG): container finished" podID="1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" containerID="bb90bf30ca9aa260b87126883ac63132095877dca04b01f21bb7ea444302ab55" exitCode=0 Nov 26 17:12:31 crc kubenswrapper[4956]: I1126 17:12:31.386422 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pdnsb" event={"ID":"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c","Type":"ContainerDied","Data":"bb90bf30ca9aa260b87126883ac63132095877dca04b01f21bb7ea444302ab55"} Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.729405 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.760746 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-config-data\") pod \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.760808 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-db-sync-config-data\") pod \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.760875 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v9bst\" (UniqueName: \"kubernetes.io/projected/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-kube-api-access-v9bst\") pod \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\" (UID: \"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c\") " Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.768088 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" (UID: "1f8ade94-6d2b-4fa9-ab08-bf3ca268757c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.770922 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-kube-api-access-v9bst" (OuterVolumeSpecName: "kube-api-access-v9bst") pod "1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" (UID: "1f8ade94-6d2b-4fa9-ab08-bf3ca268757c"). InnerVolumeSpecName "kube-api-access-v9bst". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.809967 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-config-data" (OuterVolumeSpecName: "config-data") pod "1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" (UID: "1f8ade94-6d2b-4fa9-ab08-bf3ca268757c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.862734 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.862792 4956 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:32 crc kubenswrapper[4956]: I1126 17:12:32.862898 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v9bst\" (UniqueName: \"kubernetes.io/projected/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c-kube-api-access-v9bst\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:33 crc kubenswrapper[4956]: I1126 17:12:33.424888 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-pdnsb" event={"ID":"1f8ade94-6d2b-4fa9-ab08-bf3ca268757c","Type":"ContainerDied","Data":"ac26029a37d902c1345fc8947639ae285d3a38868a6ecdfddd17c339192010b4"} Nov 26 17:12:33 crc kubenswrapper[4956]: I1126 17:12:33.425361 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ac26029a37d902c1345fc8947639ae285d3a38868a6ecdfddd17c339192010b4" Nov 26 17:12:33 crc kubenswrapper[4956]: I1126 17:12:33.424941 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-pdnsb" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.763910 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:12:34 crc kubenswrapper[4956]: E1126 17:12:34.764249 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" containerName="glance-db-sync" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.764264 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" containerName="glance-db-sync" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.764389 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" containerName="glance-db-sync" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.765424 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.768893 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.768981 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-vb95k" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.770740 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.787646 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.880309 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.881807 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.884910 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.890900 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.890969 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-scripts\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.890998 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wmjb7\" (UniqueName: \"kubernetes.io/projected/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-kube-api-access-wmjb7\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891037 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891184 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-config-data\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891239 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891308 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-run\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891348 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891435 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891493 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891568 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891599 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-dev\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891635 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-sys\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.891691 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-logs\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.896905 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992756 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992812 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-dev\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992839 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992890 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-scripts\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992912 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wmjb7\" (UniqueName: \"kubernetes.io/projected/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-kube-api-access-wmjb7\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992933 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-sys\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992951 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992956 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.992974 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993018 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-config-data\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993046 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993072 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993117 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-run\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993147 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993166 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993188 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993209 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993230 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993316 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993361 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993439 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-run\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993505 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993528 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993571 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993609 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993641 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sbf7x\" (UniqueName: \"kubernetes.io/projected/8515860e-4ae6-48a2-9a45-11b04ddd3232-kube-api-access-sbf7x\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993647 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993666 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-logs\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993704 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993728 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-dev\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993750 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993774 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-run\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993799 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-sys\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993831 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-dev\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993836 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-logs\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.993992 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.994071 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.994181 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-logs\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.994237 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-sys\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.997016 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 26 17:12:34 crc kubenswrapper[4956]: I1126 17:12:34.997201 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.008355 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-scripts\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.009088 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-config-data\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.015991 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.020271 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.028823 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wmjb7\" (UniqueName: \"kubernetes.io/projected/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-kube-api-access-wmjb7\") pod \"glance-default-external-api-0\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.082326 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-vb95k" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.091195 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095134 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-sys\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095195 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095252 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-sys\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095282 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095340 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095290 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095428 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095457 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095492 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095521 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095562 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095660 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sbf7x\" (UniqueName: \"kubernetes.io/projected/8515860e-4ae6-48a2-9a45-11b04ddd3232-kube-api-access-sbf7x\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095685 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-logs\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095725 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095748 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-run\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095821 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-dev\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095850 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.095906 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096242 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-run\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096369 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-dev\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096454 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096477 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-logs\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096618 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096806 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.096813 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.098819 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.111819 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-scripts\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.123679 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-config-data\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.128359 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.129485 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sbf7x\" (UniqueName: \"kubernetes.io/projected/8515860e-4ae6-48a2-9a45-11b04ddd3232-kube-api-access-sbf7x\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.146201 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.196796 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.644899 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:12:35 crc kubenswrapper[4956]: I1126 17:12:35.771880 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:35 crc kubenswrapper[4956]: W1126 17:12:35.800440 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8515860e_4ae6_48a2_9a45_11b04ddd3232.slice/crio-3e5947e3a25d0df15ff865b4b232f4c277153da1699cd65735a422e17ad6cb08 WatchSource:0}: Error finding container 3e5947e3a25d0df15ff865b4b232f4c277153da1699cd65735a422e17ad6cb08: Status 404 returned error can't find the container with id 3e5947e3a25d0df15ff865b4b232f4c277153da1699cd65735a422e17ad6cb08 Nov 26 17:12:36 crc kubenswrapper[4956]: I1126 17:12:36.456609 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerStarted","Data":"499706b812d257aa5178e3217a4b779fd63944df1f03c08859dd6080a135581c"} Nov 26 17:12:36 crc kubenswrapper[4956]: I1126 17:12:36.457565 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerStarted","Data":"1c0e7e13338a3866cb069f72d7ae69b9d19e7e4916c687d93f2afcb202e32576"} Nov 26 17:12:36 crc kubenswrapper[4956]: I1126 17:12:36.462729 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerStarted","Data":"ad8c541ee5bbaae02b6b1d092339219ede349a68a5ba8462b1219bb245de7aa9"} Nov 26 17:12:36 crc kubenswrapper[4956]: I1126 17:12:36.462789 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerStarted","Data":"3e5947e3a25d0df15ff865b4b232f4c277153da1699cd65735a422e17ad6cb08"} Nov 26 17:12:36 crc kubenswrapper[4956]: I1126 17:12:36.486481 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:37 crc kubenswrapper[4956]: I1126 17:12:37.480032 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerStarted","Data":"df3ad429a553cc92a4cbc319b03f06f5a92aefa78d9949729dbdbf3eb268f7be"} Nov 26 17:12:37 crc kubenswrapper[4956]: I1126 17:12:37.490953 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerStarted","Data":"7450db16bd9e52475d81036ff53fd1d1cba8bc9e378ed2be59ef9a8378e4c344"} Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.510742 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerStarted","Data":"e6dcc44d126e5420818dd9c504b4375984f99cff37117259daa7588a10d187f7"} Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.514647 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerStarted","Data":"1bfb000ec22462657d031d969989506de22d95aeaa876505b208d0614b9b19e4"} Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.514847 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-httpd" containerID="cri-o://df3ad429a553cc92a4cbc319b03f06f5a92aefa78d9949729dbdbf3eb268f7be" gracePeriod=30 Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.514824 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-log" containerID="cri-o://ad8c541ee5bbaae02b6b1d092339219ede349a68a5ba8462b1219bb245de7aa9" gracePeriod=30 Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.514847 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-api" containerID="cri-o://1bfb000ec22462657d031d969989506de22d95aeaa876505b208d0614b9b19e4" gracePeriod=30 Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.551590 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=4.551573922 podStartE2EDuration="4.551573922s" podCreationTimestamp="2025-11-26 17:12:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:38.549240556 +0000 UTC m=+1204.245201128" watchObservedRunningTime="2025-11-26 17:12:38.551573922 +0000 UTC m=+1204.247534474" Nov 26 17:12:38 crc kubenswrapper[4956]: I1126 17:12:38.585093 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=5.585067379 podStartE2EDuration="5.585067379s" podCreationTimestamp="2025-11-26 17:12:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:38.578388958 +0000 UTC m=+1204.274349560" watchObservedRunningTime="2025-11-26 17:12:38.585067379 +0000 UTC m=+1204.281027941" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.526409 4956 generic.go:334] "Generic (PLEG): container finished" podID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerID="1bfb000ec22462657d031d969989506de22d95aeaa876505b208d0614b9b19e4" exitCode=143 Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.526968 4956 generic.go:334] "Generic (PLEG): container finished" podID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerID="df3ad429a553cc92a4cbc319b03f06f5a92aefa78d9949729dbdbf3eb268f7be" exitCode=143 Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.526982 4956 generic.go:334] "Generic (PLEG): container finished" podID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerID="ad8c541ee5bbaae02b6b1d092339219ede349a68a5ba8462b1219bb245de7aa9" exitCode=143 Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.526637 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerDied","Data":"1bfb000ec22462657d031d969989506de22d95aeaa876505b208d0614b9b19e4"} Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.527390 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerDied","Data":"df3ad429a553cc92a4cbc319b03f06f5a92aefa78d9949729dbdbf3eb268f7be"} Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.527420 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerDied","Data":"ad8c541ee5bbaae02b6b1d092339219ede349a68a5ba8462b1219bb245de7aa9"} Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.762199 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957499 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-nvme\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957584 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-httpd-run\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957624 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-scripts\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957647 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-lib-modules\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957671 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-logs\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957709 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957765 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957791 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sbf7x\" (UniqueName: \"kubernetes.io/projected/8515860e-4ae6-48a2-9a45-11b04ddd3232-kube-api-access-sbf7x\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957824 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-config-data\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957885 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-iscsi\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957915 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-sys\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957938 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.957999 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-var-locks-brick\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958030 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-dev\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958054 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958073 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-run\") pod \"8515860e-4ae6-48a2-9a45-11b04ddd3232\" (UID: \"8515860e-4ae6-48a2-9a45-11b04ddd3232\") " Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958170 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958192 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-logs" (OuterVolumeSpecName: "logs") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958251 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-dev" (OuterVolumeSpecName: "dev") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958292 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-run" (OuterVolumeSpecName: "run") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958404 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958420 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958429 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8515860e-4ae6-48a2-9a45-11b04ddd3232-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958438 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958447 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958457 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958511 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958535 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.958753 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-sys" (OuterVolumeSpecName: "sys") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.964171 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-scripts" (OuterVolumeSpecName: "scripts") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.964231 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance-cache") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.964816 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:12:39 crc kubenswrapper[4956]: I1126 17:12:39.981052 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8515860e-4ae6-48a2-9a45-11b04ddd3232-kube-api-access-sbf7x" (OuterVolumeSpecName: "kube-api-access-sbf7x") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "kube-api-access-sbf7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.039262 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-config-data" (OuterVolumeSpecName: "config-data") pod "8515860e-4ae6-48a2-9a45-11b04ddd3232" (UID: "8515860e-4ae6-48a2-9a45-11b04ddd3232"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060641 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060693 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060718 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060770 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060794 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/8515860e-4ae6-48a2-9a45-11b04ddd3232-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060817 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060830 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8515860e-4ae6-48a2-9a45-11b04ddd3232-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.060843 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sbf7x\" (UniqueName: \"kubernetes.io/projected/8515860e-4ae6-48a2-9a45-11b04ddd3232-kube-api-access-sbf7x\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.075541 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.078608 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.162886 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.162932 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.539321 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"8515860e-4ae6-48a2-9a45-11b04ddd3232","Type":"ContainerDied","Data":"3e5947e3a25d0df15ff865b4b232f4c277153da1699cd65735a422e17ad6cb08"} Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.539394 4956 scope.go:117] "RemoveContainer" containerID="1bfb000ec22462657d031d969989506de22d95aeaa876505b208d0614b9b19e4" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.539552 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.584122 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.590801 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614017 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:40 crc kubenswrapper[4956]: E1126 17:12:40.614571 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-log" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614591 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-log" Nov 26 17:12:40 crc kubenswrapper[4956]: E1126 17:12:40.614610 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-httpd" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614621 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-httpd" Nov 26 17:12:40 crc kubenswrapper[4956]: E1126 17:12:40.614639 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-api" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614647 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-api" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614787 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-httpd" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614819 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-api" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.614832 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" containerName="glance-log" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.616899 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.621544 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.636600 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.772290 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.772913 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.772956 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.772991 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-dev\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773043 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvfjj\" (UniqueName: \"kubernetes.io/projected/336b8230-0428-4ee1-bc0c-bab3369d43e1-kube-api-access-dvfjj\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773105 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-sys\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773146 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773185 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773216 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773248 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773287 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773319 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-run\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773349 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.773393 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.857549 4956 scope.go:117] "RemoveContainer" containerID="df3ad429a553cc92a4cbc319b03f06f5a92aefa78d9949729dbdbf3eb268f7be" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874477 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-sys\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874537 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874563 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874584 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-sys\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874663 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874696 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874753 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874770 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874789 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-run\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875091 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874971 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874998 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") device mount path \"/mnt/openstack/pv15\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875223 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875265 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875308 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.874999 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-run\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875008 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875435 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875520 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875560 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875584 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-dev\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875614 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvfjj\" (UniqueName: \"kubernetes.io/projected/336b8230-0428-4ee1-bc0c-bab3369d43e1-kube-api-access-dvfjj\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875810 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875938 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-dev\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.875975 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-logs\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.878576 4956 scope.go:117] "RemoveContainer" containerID="ad8c541ee5bbaae02b6b1d092339219ede349a68a5ba8462b1219bb245de7aa9" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.880198 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-scripts\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.882926 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-config-data\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.895488 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvfjj\" (UniqueName: \"kubernetes.io/projected/336b8230-0428-4ee1-bc0c-bab3369d43e1-kube-api-access-dvfjj\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.900996 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.901499 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"glance-default-internal-api-0\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:40 crc kubenswrapper[4956]: I1126 17:12:40.945411 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:41 crc kubenswrapper[4956]: I1126 17:12:41.018328 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8515860e-4ae6-48a2-9a45-11b04ddd3232" path="/var/lib/kubelet/pods/8515860e-4ae6-48a2-9a45-11b04ddd3232/volumes" Nov 26 17:12:41 crc kubenswrapper[4956]: I1126 17:12:41.441021 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:12:41 crc kubenswrapper[4956]: I1126 17:12:41.549062 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerStarted","Data":"000e20dc49dcdfbd6a08be8365827fa5c53d2e91822e553aeda61de19b05b8cf"} Nov 26 17:12:42 crc kubenswrapper[4956]: I1126 17:12:42.576827 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerStarted","Data":"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549"} Nov 26 17:12:42 crc kubenswrapper[4956]: I1126 17:12:42.577593 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerStarted","Data":"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d"} Nov 26 17:12:42 crc kubenswrapper[4956]: I1126 17:12:42.577628 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerStarted","Data":"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f"} Nov 26 17:12:42 crc kubenswrapper[4956]: I1126 17:12:42.601650 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.6016285420000003 podStartE2EDuration="2.601628542s" podCreationTimestamp="2025-11-26 17:12:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:42.600676335 +0000 UTC m=+1208.296636877" watchObservedRunningTime="2025-11-26 17:12:42.601628542 +0000 UTC m=+1208.297589094" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.092967 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.093325 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.093337 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.125919 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.136743 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.137814 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.619064 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.619124 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.619142 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.635323 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.639492 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:45 crc kubenswrapper[4956]: I1126 17:12:45.639745 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:12:50 crc kubenswrapper[4956]: I1126 17:12:50.946163 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:50 crc kubenswrapper[4956]: I1126 17:12:50.946770 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:50 crc kubenswrapper[4956]: I1126 17:12:50.946787 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:50 crc kubenswrapper[4956]: I1126 17:12:50.978513 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:50 crc kubenswrapper[4956]: I1126 17:12:50.980085 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.012169 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.680194 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.680250 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.680261 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.693845 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.695375 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:51 crc kubenswrapper[4956]: I1126 17:12:51.699107 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.201986 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.205939 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.213991 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.216398 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.231299 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.237916 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.351592 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.353961 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.360618 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.360679 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-scripts\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.360816 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r64p5\" (UniqueName: \"kubernetes.io/projected/9cb132c4-0419-4520-ba16-e5f40377e318-kube-api-access-r64p5\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.360898 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.360952 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-sys\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361009 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361069 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-config-data\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361093 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361117 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-run\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361139 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-sys\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361190 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-run\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361223 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-config-data\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361250 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-scripts\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361274 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fktx4\" (UniqueName: \"kubernetes.io/projected/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-kube-api-access-fktx4\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361307 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-dev\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361363 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361464 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-dev\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361507 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361537 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.361854 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-logs\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362014 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362061 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362108 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-logs\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362163 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362184 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362267 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362285 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.362342 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.365932 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.368087 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.387962 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.399022 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.464067 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.464906 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-config-data\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465065 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-config-data\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465249 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-dev\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465375 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465540 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465651 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465790 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-logs\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.465979 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-dev\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466113 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466132 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466386 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-run\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466486 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-27m78\" (UniqueName: \"kubernetes.io/projected/7df26fee-12cb-4d53-a0d1-b33a106f141c-kube-api-access-27m78\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466612 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466707 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466802 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.466933 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-scripts\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467040 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467141 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467255 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gs6vn\" (UniqueName: \"kubernetes.io/projected/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-kube-api-access-gs6vn\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467381 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-logs\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467502 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467606 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467715 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467827 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.467943 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468149 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468304 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468458 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468461 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468531 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468579 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-sys\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468610 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468632 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468666 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-logs\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468706 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468740 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468791 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-run\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468837 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-scripts\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468885 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r64p5\" (UniqueName: \"kubernetes.io/projected/9cb132c4-0419-4520-ba16-e5f40377e318-kube-api-access-r64p5\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468908 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468931 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468962 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-sys\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.468993 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-dev\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469015 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469045 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-sys\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469071 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-scripts\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469100 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-config-data\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469118 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469216 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-run\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469265 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-sys\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469316 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-dev\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469359 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-run\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469387 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469425 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-config-data\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469451 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-scripts\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469479 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fktx4\" (UniqueName: \"kubernetes.io/projected/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-kube-api-access-fktx4\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469514 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-dev\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469545 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469568 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.469618 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-logs\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.470551 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-logs\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.470617 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-run\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.470656 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-sys\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.470705 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-run\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.470777 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.471034 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.472705 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.473138 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-logs\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.473377 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.474263 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.475525 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-sys\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.474620 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.474716 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.475030 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.474764 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.474579 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.474745 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-dev\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.483265 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-scripts\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.516458 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-scripts\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.523952 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-config-data\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.524391 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-config-data\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.529119 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r64p5\" (UniqueName: \"kubernetes.io/projected/9cb132c4-0419-4520-ba16-e5f40377e318-kube-api-access-r64p5\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.529743 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fktx4\" (UniqueName: \"kubernetes.io/projected/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-kube-api-access-fktx4\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.568084 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.568561 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571458 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-dev\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571503 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-sys\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571528 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-scripts\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571558 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-dev\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571577 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571621 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-logs\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571642 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571660 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-config-data\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571679 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-config-data\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571702 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571721 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-run\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571741 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-27m78\" (UniqueName: \"kubernetes.io/projected/7df26fee-12cb-4d53-a0d1-b33a106f141c-kube-api-access-27m78\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571756 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571772 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571793 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-scripts\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571810 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571831 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gs6vn\" (UniqueName: \"kubernetes.io/projected/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-kube-api-access-gs6vn\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571850 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571885 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571906 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571926 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571940 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571965 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-sys\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.571982 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572001 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-logs\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572025 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572048 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-run\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572080 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572184 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572217 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-dev\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572236 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-sys\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.572983 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573164 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573291 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573281 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573832 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573340 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573365 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-run\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573365 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-run\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573403 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.574054 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-dev\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.574064 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573517 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.574094 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573447 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.574343 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-logs\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.574434 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573339 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-sys\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573750 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-logs\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.573435 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.576422 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-scripts\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.578974 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-config-data\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.581257 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-external-api-1\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.588568 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-scripts\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.591243 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-config-data\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.594500 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-27m78\" (UniqueName: \"kubernetes.io/projected/7df26fee-12cb-4d53-a0d1-b33a106f141c-kube-api-access-27m78\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.598754 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.626298 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gs6vn\" (UniqueName: \"kubernetes.io/projected/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-kube-api-access-gs6vn\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.633208 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.649130 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.649800 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-2\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.650221 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-2\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.654376 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-1\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.676184 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.694525 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.867996 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:12:54 crc kubenswrapper[4956]: W1126 17:12:54.874209 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb20c7ec5_2a9f_4a61_ad51_c09d5212a0a4.slice/crio-2269a4d6f92b49fd8eede718d003b35069d5806b49a20a3c8341deaced1659c3 WatchSource:0}: Error finding container 2269a4d6f92b49fd8eede718d003b35069d5806b49a20a3c8341deaced1659c3: Status 404 returned error can't find the container with id 2269a4d6f92b49fd8eede718d003b35069d5806b49a20a3c8341deaced1659c3 Nov 26 17:12:54 crc kubenswrapper[4956]: I1126 17:12:54.912543 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.188646 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:12:55 crc kubenswrapper[4956]: W1126 17:12:55.196718 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9cb132c4_0419_4520_ba16_e5f40377e318.slice/crio-273eae65b50fc391f58036a259069b9592048ef9c51fcec93d6f87ac4cb1468b WatchSource:0}: Error finding container 273eae65b50fc391f58036a259069b9592048ef9c51fcec93d6f87ac4cb1468b: Status 404 returned error can't find the container with id 273eae65b50fc391f58036a259069b9592048ef9c51fcec93d6f87ac4cb1468b Nov 26 17:12:55 crc kubenswrapper[4956]: W1126 17:12:55.200580 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7df26fee_12cb_4d53_a0d1_b33a106f141c.slice/crio-4a2f737916ce68ff749fc1d7023c42d004a17c7bb60134996361fb7c6d7b7b87 WatchSource:0}: Error finding container 4a2f737916ce68ff749fc1d7023c42d004a17c7bb60134996361fb7c6d7b7b87: Status 404 returned error can't find the container with id 4a2f737916ce68ff749fc1d7023c42d004a17c7bb60134996361fb7c6d7b7b87 Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.202612 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.245913 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.731323 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerStarted","Data":"309c0cfc827a22704d8edbe4e87a43a8fda55c32bafae2a2aecc234bd6502b8e"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.731756 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerStarted","Data":"5d76db8c6b605ea63ff82ffe811c91ae24aa9ef5cb188795873a283343a07989"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.731774 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerStarted","Data":"d4f01a6a0abeef808dff300afc847a079c2c246201d586530d2478c527344ad7"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.731801 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerStarted","Data":"2269a4d6f92b49fd8eede718d003b35069d5806b49a20a3c8341deaced1659c3"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.738273 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerStarted","Data":"6247e2c04e8ac86e9d6fc6821fd9c05691a2d7b524e6f72a38794960f71a954b"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.738920 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerStarted","Data":"4a2f737916ce68ff749fc1d7023c42d004a17c7bb60134996361fb7c6d7b7b87"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.741806 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerStarted","Data":"19d2f0b7cc19e07898c1bc918bd57d41c11ab0ea4534627e9bed3bd426253552"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.741968 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerStarted","Data":"273eae65b50fc391f58036a259069b9592048ef9c51fcec93d6f87ac4cb1468b"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.755029 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerStarted","Data":"2ed7720c14ba9388e816f0b21cf5dec8711b468ec784894b7f2edf0a0f51790e"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.755151 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerStarted","Data":"5cbd7e4cf2b6e577666df1f99e9cccb69a230d28beb590eec015fe4f30f5cd9c"} Nov 26 17:12:55 crc kubenswrapper[4956]: I1126 17:12:55.787648 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.787625121 podStartE2EDuration="2.787625121s" podCreationTimestamp="2025-11-26 17:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:55.779756427 +0000 UTC m=+1221.475716979" watchObservedRunningTime="2025-11-26 17:12:55.787625121 +0000 UTC m=+1221.483585693" Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.766997 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerStarted","Data":"66d63671d43215573e0b0aca41878a41fb2d3c9086568d2b4de8c81d7c456406"} Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.767706 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerStarted","Data":"f00bcdcb31c0d21691a92b240d21ee8cf8b53a027b50f4b4619cc64b4ad4414f"} Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.770205 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerStarted","Data":"3bc97cae5660442a4b69301f5a0acca727ea86da6665aa6c6c1affbbbc435013"} Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.770274 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerStarted","Data":"f5fb15f7c20bf7483cf299fab0e7598392a281a91cd65caf40ed7978715bf356"} Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.776288 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerStarted","Data":"e311831f796382aee1909eb34f5137668e45523562cd84ad6787f37a9bf4e053"} Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.776350 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerStarted","Data":"fa5d19e21037c68f743e370c27692ad2a1f2921dd593e225e04ed33064ead351"} Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.798237 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.79820514 podStartE2EDuration="3.79820514s" podCreationTimestamp="2025-11-26 17:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:56.792220499 +0000 UTC m=+1222.488181051" watchObservedRunningTime="2025-11-26 17:12:56.79820514 +0000 UTC m=+1222.494165692" Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.826729 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.826704065 podStartE2EDuration="3.826704065s" podCreationTimestamp="2025-11-26 17:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:56.818451629 +0000 UTC m=+1222.514412191" watchObservedRunningTime="2025-11-26 17:12:56.826704065 +0000 UTC m=+1222.522664627" Nov 26 17:12:56 crc kubenswrapper[4956]: I1126 17:12:56.854472 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=3.854450217 podStartE2EDuration="3.854450217s" podCreationTimestamp="2025-11-26 17:12:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:12:56.848758595 +0000 UTC m=+1222.544719147" watchObservedRunningTime="2025-11-26 17:12:56.854450217 +0000 UTC m=+1222.550410759" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.599340 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.599817 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.599827 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.625547 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.627685 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.639116 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.676372 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.676422 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.676431 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.695059 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.695121 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.695134 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.706605 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.716170 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.728644 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.728766 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.731413 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.742430 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.844284 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.844360 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845034 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845094 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845110 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845126 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845139 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845149 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.845162 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.857478 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.860955 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.861274 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.862995 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.863041 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.863287 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.863786 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.866364 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.867752 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.915355 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.917778 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.917901 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.961296 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.963915 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:04 crc kubenswrapper[4956]: I1126 17:13:04.989500 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:05 crc kubenswrapper[4956]: I1126 17:13:05.852913 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:05 crc kubenswrapper[4956]: I1126 17:13:05.852961 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:05 crc kubenswrapper[4956]: I1126 17:13:05.852974 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:05 crc kubenswrapper[4956]: I1126 17:13:05.868562 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:05 crc kubenswrapper[4956]: I1126 17:13:05.869396 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:05 crc kubenswrapper[4956]: I1126 17:13:05.878816 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:06 crc kubenswrapper[4956]: I1126 17:13:06.824788 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:13:06 crc kubenswrapper[4956]: I1126 17:13:06.838298 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.175637 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.185278 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.883435 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-api" containerID="cri-o://e311831f796382aee1909eb34f5137668e45523562cd84ad6787f37a9bf4e053" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.883308 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-log" containerID="cri-o://2ed7720c14ba9388e816f0b21cf5dec8711b468ec784894b7f2edf0a0f51790e" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.883668 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-log" containerID="cri-o://6247e2c04e8ac86e9d6fc6821fd9c05691a2d7b524e6f72a38794960f71a954b" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.885220 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-api" containerID="cri-o://3bc97cae5660442a4b69301f5a0acca727ea86da6665aa6c6c1affbbbc435013" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.885343 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-httpd" containerID="cri-o://f5fb15f7c20bf7483cf299fab0e7598392a281a91cd65caf40ed7978715bf356" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.883476 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-httpd" containerID="cri-o://fa5d19e21037c68f743e370c27692ad2a1f2921dd593e225e04ed33064ead351" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.884473 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-log" containerID="cri-o://19d2f0b7cc19e07898c1bc918bd57d41c11ab0ea4534627e9bed3bd426253552" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.885027 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-log" containerID="cri-o://d4f01a6a0abeef808dff300afc847a079c2c246201d586530d2478c527344ad7" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.883739 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-api" containerID="cri-o://66d63671d43215573e0b0aca41878a41fb2d3c9086568d2b4de8c81d7c456406" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.885590 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-api" containerID="cri-o://309c0cfc827a22704d8edbe4e87a43a8fda55c32bafae2a2aecc234bd6502b8e" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.885631 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-httpd" containerID="cri-o://5d76db8c6b605ea63ff82ffe811c91ae24aa9ef5cb188795873a283343a07989" gracePeriod=30 Nov 26 17:13:07 crc kubenswrapper[4956]: I1126 17:13:07.883757 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-httpd" containerID="cri-o://f00bcdcb31c0d21691a92b240d21ee8cf8b53a027b50f4b4619cc64b4ad4414f" gracePeriod=30 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.902816 4956 generic.go:334] "Generic (PLEG): container finished" podID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerID="66d63671d43215573e0b0aca41878a41fb2d3c9086568d2b4de8c81d7c456406" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.903133 4956 generic.go:334] "Generic (PLEG): container finished" podID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerID="f00bcdcb31c0d21691a92b240d21ee8cf8b53a027b50f4b4619cc64b4ad4414f" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.903143 4956 generic.go:334] "Generic (PLEG): container finished" podID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerID="6247e2c04e8ac86e9d6fc6821fd9c05691a2d7b524e6f72a38794960f71a954b" exitCode=143 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.903006 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerDied","Data":"66d63671d43215573e0b0aca41878a41fb2d3c9086568d2b4de8c81d7c456406"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.903224 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerDied","Data":"f00bcdcb31c0d21691a92b240d21ee8cf8b53a027b50f4b4619cc64b4ad4414f"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.903239 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerDied","Data":"6247e2c04e8ac86e9d6fc6821fd9c05691a2d7b524e6f72a38794960f71a954b"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.905416 4956 generic.go:334] "Generic (PLEG): container finished" podID="9cb132c4-0419-4520-ba16-e5f40377e318" containerID="3bc97cae5660442a4b69301f5a0acca727ea86da6665aa6c6c1affbbbc435013" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.905434 4956 generic.go:334] "Generic (PLEG): container finished" podID="9cb132c4-0419-4520-ba16-e5f40377e318" containerID="f5fb15f7c20bf7483cf299fab0e7598392a281a91cd65caf40ed7978715bf356" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.905444 4956 generic.go:334] "Generic (PLEG): container finished" podID="9cb132c4-0419-4520-ba16-e5f40377e318" containerID="19d2f0b7cc19e07898c1bc918bd57d41c11ab0ea4534627e9bed3bd426253552" exitCode=143 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.905470 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerDied","Data":"3bc97cae5660442a4b69301f5a0acca727ea86da6665aa6c6c1affbbbc435013"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.905486 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerDied","Data":"f5fb15f7c20bf7483cf299fab0e7598392a281a91cd65caf40ed7978715bf356"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.905495 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerDied","Data":"19d2f0b7cc19e07898c1bc918bd57d41c11ab0ea4534627e9bed3bd426253552"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.911086 4956 generic.go:334] "Generic (PLEG): container finished" podID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerID="e311831f796382aee1909eb34f5137668e45523562cd84ad6787f37a9bf4e053" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.911106 4956 generic.go:334] "Generic (PLEG): container finished" podID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerID="fa5d19e21037c68f743e370c27692ad2a1f2921dd593e225e04ed33064ead351" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.911115 4956 generic.go:334] "Generic (PLEG): container finished" podID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerID="2ed7720c14ba9388e816f0b21cf5dec8711b468ec784894b7f2edf0a0f51790e" exitCode=143 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.911183 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerDied","Data":"e311831f796382aee1909eb34f5137668e45523562cd84ad6787f37a9bf4e053"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.911243 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerDied","Data":"fa5d19e21037c68f743e370c27692ad2a1f2921dd593e225e04ed33064ead351"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.911259 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerDied","Data":"2ed7720c14ba9388e816f0b21cf5dec8711b468ec784894b7f2edf0a0f51790e"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.914164 4956 generic.go:334] "Generic (PLEG): container finished" podID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerID="309c0cfc827a22704d8edbe4e87a43a8fda55c32bafae2a2aecc234bd6502b8e" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.914207 4956 generic.go:334] "Generic (PLEG): container finished" podID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerID="5d76db8c6b605ea63ff82ffe811c91ae24aa9ef5cb188795873a283343a07989" exitCode=0 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.914201 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerDied","Data":"309c0cfc827a22704d8edbe4e87a43a8fda55c32bafae2a2aecc234bd6502b8e"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.914261 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerDied","Data":"5d76db8c6b605ea63ff82ffe811c91ae24aa9ef5cb188795873a283343a07989"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.914274 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerDied","Data":"d4f01a6a0abeef808dff300afc847a079c2c246201d586530d2478c527344ad7"} Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.914218 4956 generic.go:334] "Generic (PLEG): container finished" podID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerID="d4f01a6a0abeef808dff300afc847a079c2c246201d586530d2478c527344ad7" exitCode=143 Nov 26 17:13:08 crc kubenswrapper[4956]: I1126 17:13:08.973741 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064379 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gs6vn\" (UniqueName: \"kubernetes.io/projected/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-kube-api-access-gs6vn\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064428 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-config-data\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064469 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-iscsi\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064498 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064543 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-dev\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064651 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-var-locks-brick\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064690 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-run\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064747 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-lib-modules\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064769 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-nvme\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064816 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-logs\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.064905 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-scripts\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.065006 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.065032 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-sys\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.065072 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-httpd-run\") pod \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\" (UID: \"ea8552e9-d8c9-4862-8d7b-9b019f8d4524\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.070180 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.070286 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.071849 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.071922 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-dev" (OuterVolumeSpecName: "dev") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.072257 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.072523 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-logs" (OuterVolumeSpecName: "logs") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.072547 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-sys" (OuterVolumeSpecName: "sys") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.073132 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.073302 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-run" (OuterVolumeSpecName: "run") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.073505 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-kube-api-access-gs6vn" (OuterVolumeSpecName: "kube-api-access-gs6vn") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "kube-api-access-gs6vn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.074599 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-scripts" (OuterVolumeSpecName: "scripts") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.076763 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.096710 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168060 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gs6vn\" (UniqueName: \"kubernetes.io/projected/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-kube-api-access-gs6vn\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168099 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168142 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168160 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168172 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168185 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168196 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168206 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168216 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168227 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168253 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168266 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.168277 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.172990 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-config-data" (OuterVolumeSpecName: "config-data") pod "ea8552e9-d8c9-4862-8d7b-9b019f8d4524" (UID: "ea8552e9-d8c9-4862-8d7b-9b019f8d4524"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.198941 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.201221 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.270295 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.270334 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ea8552e9-d8c9-4862-8d7b-9b019f8d4524-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.270345 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.340125 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.341326 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.355492 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371612 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-nvme\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371659 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-logs\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371683 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-sys\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371705 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371719 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-run\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371736 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-httpd-run\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371749 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-var-locks-brick\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371770 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-config-data\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371794 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-sys\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371816 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fktx4\" (UniqueName: \"kubernetes.io/projected/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-kube-api-access-fktx4\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371837 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-lib-modules\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371856 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-logs\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371899 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-run\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371921 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-nvme\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371945 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-httpd-run\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371963 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-iscsi\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.371980 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-run\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372003 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r64p5\" (UniqueName: \"kubernetes.io/projected/9cb132c4-0419-4520-ba16-e5f40377e318-kube-api-access-r64p5\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372019 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-logs\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372044 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372083 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-config-data\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372115 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-lib-modules\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372144 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-dev\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372159 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-scripts\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372177 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-httpd-run\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372192 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-dev\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372208 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-nvme\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372222 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-lib-modules\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372248 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372272 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-config-data\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372287 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-iscsi\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372308 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372324 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-scripts\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372341 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372359 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-dev\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372380 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-sys\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372394 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-iscsi\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372408 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-var-locks-brick\") pod \"9cb132c4-0419-4520-ba16-e5f40377e318\" (UID: \"9cb132c4-0419-4520-ba16-e5f40377e318\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372426 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-27m78\" (UniqueName: \"kubernetes.io/projected/7df26fee-12cb-4d53-a0d1-b33a106f141c-kube-api-access-27m78\") pod \"7df26fee-12cb-4d53-a0d1-b33a106f141c\" (UID: \"7df26fee-12cb-4d53-a0d1-b33a106f141c\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372447 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-scripts\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372466 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372485 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-var-locks-brick\") pod \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\" (UID: \"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4\") " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372811 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372843 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372883 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.372906 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-dev" (OuterVolumeSpecName: "dev") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.376793 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.377213 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.377918 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-logs" (OuterVolumeSpecName: "logs") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.377955 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-sys" (OuterVolumeSpecName: "sys") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.378950 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-scripts" (OuterVolumeSpecName: "scripts") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.379019 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-scripts" (OuterVolumeSpecName: "scripts") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.379091 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.379202 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.379238 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-dev" (OuterVolumeSpecName: "dev") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.379262 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.379282 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380352 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance-cache") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380364 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380412 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-dev" (OuterVolumeSpecName: "dev") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380439 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-run" (OuterVolumeSpecName: "run") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380440 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-sys" (OuterVolumeSpecName: "sys") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380480 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-sys" (OuterVolumeSpecName: "sys") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.380686 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381488 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381524 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381599 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance-cache") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381654 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-run" (OuterVolumeSpecName: "run") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381682 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381714 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-run" (OuterVolumeSpecName: "run") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.381988 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-logs" (OuterVolumeSpecName: "logs") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.382057 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-logs" (OuterVolumeSpecName: "logs") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.382434 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.382682 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.384782 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7df26fee-12cb-4d53-a0d1-b33a106f141c-kube-api-access-27m78" (OuterVolumeSpecName: "kube-api-access-27m78") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "kube-api-access-27m78". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.385063 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.385204 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.388228 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9cb132c4-0419-4520-ba16-e5f40377e318-kube-api-access-r64p5" (OuterVolumeSpecName: "kube-api-access-r64p5") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "kube-api-access-r64p5". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.388342 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-scripts" (OuterVolumeSpecName: "scripts") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.392110 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-kube-api-access-fktx4" (OuterVolumeSpecName: "kube-api-access-fktx4") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "kube-api-access-fktx4". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.394067 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474057 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474094 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474107 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474116 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474126 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fktx4\" (UniqueName: \"kubernetes.io/projected/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-kube-api-access-fktx4\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474138 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474146 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474156 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474165 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474177 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474188 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474197 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474205 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r64p5\" (UniqueName: \"kubernetes.io/projected/9cb132c4-0419-4520-ba16-e5f40377e318-kube-api-access-r64p5\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474213 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7df26fee-12cb-4d53-a0d1-b33a106f141c-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474241 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474250 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474259 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474267 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474275 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/9cb132c4-0419-4520-ba16-e5f40377e318-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474283 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474292 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474300 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474313 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474322 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474336 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474344 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474357 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474368 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474376 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474384 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474395 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/9cb132c4-0419-4520-ba16-e5f40377e318-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474404 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-27m78\" (UniqueName: \"kubernetes.io/projected/7df26fee-12cb-4d53-a0d1-b33a106f141c-kube-api-access-27m78\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474415 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474427 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474437 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474447 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7df26fee-12cb-4d53-a0d1-b33a106f141c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474457 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474470 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.474485 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.481978 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-config-data" (OuterVolumeSpecName: "config-data") pod "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" (UID: "b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.488620 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.488929 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.490111 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.491299 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.497057 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.497334 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.507642 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-config-data" (OuterVolumeSpecName: "config-data") pod "9cb132c4-0419-4520-ba16-e5f40377e318" (UID: "9cb132c4-0419-4520-ba16-e5f40377e318"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.507962 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-config-data" (OuterVolumeSpecName: "config-data") pod "7df26fee-12cb-4d53-a0d1-b33a106f141c" (UID: "7df26fee-12cb-4d53-a0d1-b33a106f141c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577053 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577109 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9cb132c4-0419-4520-ba16-e5f40377e318-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577125 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577163 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577178 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577193 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577207 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7df26fee-12cb-4d53-a0d1-b33a106f141c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577219 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.577232 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.932556 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"7df26fee-12cb-4d53-a0d1-b33a106f141c","Type":"ContainerDied","Data":"4a2f737916ce68ff749fc1d7023c42d004a17c7bb60134996361fb7c6d7b7b87"} Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.932667 4956 scope.go:117] "RemoveContainer" containerID="66d63671d43215573e0b0aca41878a41fb2d3c9086568d2b4de8c81d7c456406" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.933130 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.940376 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"9cb132c4-0419-4520-ba16-e5f40377e318","Type":"ContainerDied","Data":"273eae65b50fc391f58036a259069b9592048ef9c51fcec93d6f87ac4cb1468b"} Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.941681 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.963660 4956 scope.go:117] "RemoveContainer" containerID="f00bcdcb31c0d21691a92b240d21ee8cf8b53a027b50f4b4619cc64b4ad4414f" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.966208 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.968109 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"ea8552e9-d8c9-4862-8d7b-9b019f8d4524","Type":"ContainerDied","Data":"5cbd7e4cf2b6e577666df1f99e9cccb69a230d28beb590eec015fe4f30f5cd9c"} Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.970998 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4","Type":"ContainerDied","Data":"2269a4d6f92b49fd8eede718d003b35069d5806b49a20a3c8341deaced1659c3"} Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.971112 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:09 crc kubenswrapper[4956]: I1126 17:13:09.997182 4956 scope.go:117] "RemoveContainer" containerID="6247e2c04e8ac86e9d6fc6821fd9c05691a2d7b524e6f72a38794960f71a954b" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.005421 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.011253 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.031333 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.035584 4956 scope.go:117] "RemoveContainer" containerID="3bc97cae5660442a4b69301f5a0acca727ea86da6665aa6c6c1affbbbc435013" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.039375 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.049943 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.053446 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.066312 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.077883 4956 scope.go:117] "RemoveContainer" containerID="f5fb15f7c20bf7483cf299fab0e7598392a281a91cd65caf40ed7978715bf356" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.081292 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.100658 4956 scope.go:117] "RemoveContainer" containerID="19d2f0b7cc19e07898c1bc918bd57d41c11ab0ea4534627e9bed3bd426253552" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.123973 4956 scope.go:117] "RemoveContainer" containerID="e311831f796382aee1909eb34f5137668e45523562cd84ad6787f37a9bf4e053" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.154275 4956 scope.go:117] "RemoveContainer" containerID="fa5d19e21037c68f743e370c27692ad2a1f2921dd593e225e04ed33064ead351" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.175632 4956 scope.go:117] "RemoveContainer" containerID="2ed7720c14ba9388e816f0b21cf5dec8711b468ec784894b7f2edf0a0f51790e" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.205301 4956 scope.go:117] "RemoveContainer" containerID="309c0cfc827a22704d8edbe4e87a43a8fda55c32bafae2a2aecc234bd6502b8e" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.253421 4956 scope.go:117] "RemoveContainer" containerID="5d76db8c6b605ea63ff82ffe811c91ae24aa9ef5cb188795873a283343a07989" Nov 26 17:13:10 crc kubenswrapper[4956]: I1126 17:13:10.272752 4956 scope.go:117] "RemoveContainer" containerID="d4f01a6a0abeef808dff300afc847a079c2c246201d586530d2478c527344ad7" Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.005563 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" path="/var/lib/kubelet/pods/7df26fee-12cb-4d53-a0d1-b33a106f141c/volumes" Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.006576 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" path="/var/lib/kubelet/pods/9cb132c4-0419-4520-ba16-e5f40377e318/volumes" Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.007822 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" path="/var/lib/kubelet/pods/b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4/volumes" Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.008648 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" path="/var/lib/kubelet/pods/ea8552e9-d8c9-4862-8d7b-9b019f8d4524/volumes" Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.117007 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.117417 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-log" containerID="cri-o://499706b812d257aa5178e3217a4b779fd63944df1f03c08859dd6080a135581c" gracePeriod=30 Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.117550 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-api" containerID="cri-o://e6dcc44d126e5420818dd9c504b4375984f99cff37117259daa7588a10d187f7" gracePeriod=30 Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.117585 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-httpd" containerID="cri-o://7450db16bd9e52475d81036ff53fd1d1cba8bc9e378ed2be59ef9a8378e4c344" gracePeriod=30 Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.774596 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.775273 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-log" containerID="cri-o://5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f" gracePeriod=30 Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.775330 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-api" containerID="cri-o://57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549" gracePeriod=30 Nov 26 17:13:11 crc kubenswrapper[4956]: I1126 17:13:11.775405 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-httpd" containerID="cri-o://814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d" gracePeriod=30 Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.011705 4956 generic.go:334] "Generic (PLEG): container finished" podID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerID="5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f" exitCode=143 Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.011792 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerDied","Data":"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f"} Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.014918 4956 generic.go:334] "Generic (PLEG): container finished" podID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerID="e6dcc44d126e5420818dd9c504b4375984f99cff37117259daa7588a10d187f7" exitCode=0 Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.014944 4956 generic.go:334] "Generic (PLEG): container finished" podID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerID="7450db16bd9e52475d81036ff53fd1d1cba8bc9e378ed2be59ef9a8378e4c344" exitCode=0 Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.014953 4956 generic.go:334] "Generic (PLEG): container finished" podID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerID="499706b812d257aa5178e3217a4b779fd63944df1f03c08859dd6080a135581c" exitCode=143 Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.014972 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerDied","Data":"e6dcc44d126e5420818dd9c504b4375984f99cff37117259daa7588a10d187f7"} Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.014998 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerDied","Data":"7450db16bd9e52475d81036ff53fd1d1cba8bc9e378ed2be59ef9a8378e4c344"} Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.015008 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerDied","Data":"499706b812d257aa5178e3217a4b779fd63944df1f03c08859dd6080a135581c"} Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.037119 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122553 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-dev\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122635 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-config-data\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122667 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-nvme\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122694 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-scripts\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122713 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-iscsi\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122723 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-dev" (OuterVolumeSpecName: "dev") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122778 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wmjb7\" (UniqueName: \"kubernetes.io/projected/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-kube-api-access-wmjb7\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122805 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122830 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122919 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-logs\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.122981 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-var-locks-brick\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123002 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123044 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-httpd-run\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123083 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-sys\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123110 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-run\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123156 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-lib-modules\") pod \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\" (UID: \"4cbc07d2-febf-433f-87ed-4bcdff10b3f8\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123736 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123757 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.123796 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.124101 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.124128 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-sys" (OuterVolumeSpecName: "sys") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.124148 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-run" (OuterVolumeSpecName: "run") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.124172 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.124262 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.124520 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-logs" (OuterVolumeSpecName: "logs") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.129383 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-scripts" (OuterVolumeSpecName: "scripts") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.129410 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.132281 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.135624 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-kube-api-access-wmjb7" (OuterVolumeSpecName: "kube-api-access-wmjb7") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "kube-api-access-wmjb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.199132 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-config-data" (OuterVolumeSpecName: "config-data") pod "4cbc07d2-febf-433f-87ed-4bcdff10b3f8" (UID: "4cbc07d2-febf-433f-87ed-4bcdff10b3f8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225112 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225166 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225179 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225194 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225203 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225212 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225220 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225228 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225237 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wmjb7\" (UniqueName: \"kubernetes.io/projected/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-kube-api-access-wmjb7\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225268 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225276 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.225285 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4cbc07d2-febf-433f-87ed-4bcdff10b3f8-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.238687 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.246826 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.327333 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.327391 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.832392 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.936940 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvfjj\" (UniqueName: \"kubernetes.io/projected/336b8230-0428-4ee1-bc0c-bab3369d43e1-kube-api-access-dvfjj\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937332 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937364 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-httpd-run\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937414 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-iscsi\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937435 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937592 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937630 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-run\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937693 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-var-locks-brick\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937748 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-sys\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937784 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-config-data\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937808 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-scripts\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937827 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-nvme\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937938 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-logs\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937962 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-dev\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.937991 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-lib-modules\") pod \"336b8230-0428-4ee1-bc0c-bab3369d43e1\" (UID: \"336b8230-0428-4ee1-bc0c-bab3369d43e1\") " Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938003 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938388 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938406 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938440 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938567 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938631 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-run" (OuterVolumeSpecName: "run") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938665 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-dev" (OuterVolumeSpecName: "dev") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938701 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.938735 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-sys" (OuterVolumeSpecName: "sys") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.939045 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-logs" (OuterVolumeSpecName: "logs") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.941220 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.942338 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage15-crc" (OuterVolumeSpecName: "glance-cache") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "local-storage15-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.942905 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/336b8230-0428-4ee1-bc0c-bab3369d43e1-kube-api-access-dvfjj" (OuterVolumeSpecName: "kube-api-access-dvfjj") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "kube-api-access-dvfjj". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:12 crc kubenswrapper[4956]: I1126 17:13:12.943355 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-scripts" (OuterVolumeSpecName: "scripts") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.023982 4956 generic.go:334] "Generic (PLEG): container finished" podID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerID="57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549" exitCode=0 Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.024017 4956 generic.go:334] "Generic (PLEG): container finished" podID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerID="814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d" exitCode=0 Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.024062 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.024068 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerDied","Data":"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549"} Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.024097 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerDied","Data":"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d"} Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.024107 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"336b8230-0428-4ee1-bc0c-bab3369d43e1","Type":"ContainerDied","Data":"000e20dc49dcdfbd6a08be8365827fa5c53d2e91822e553aeda61de19b05b8cf"} Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.024124 4956 scope.go:117] "RemoveContainer" containerID="57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.025035 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-config-data" (OuterVolumeSpecName: "config-data") pod "336b8230-0428-4ee1-bc0c-bab3369d43e1" (UID: "336b8230-0428-4ee1-bc0c-bab3369d43e1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.026415 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4cbc07d2-febf-433f-87ed-4bcdff10b3f8","Type":"ContainerDied","Data":"1c0e7e13338a3866cb069f72d7ae69b9d19e7e4916c687d93f2afcb202e32576"} Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.026496 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039788 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvfjj\" (UniqueName: \"kubernetes.io/projected/336b8230-0428-4ee1-bc0c-bab3369d43e1-kube-api-access-dvfjj\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039837 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" " Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039852 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039905 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039923 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039935 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039943 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039953 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/336b8230-0428-4ee1-bc0c-bab3369d43e1-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039961 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039969 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/336b8230-0428-4ee1-bc0c-bab3369d43e1-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039978 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.039987 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/336b8230-0428-4ee1-bc0c-bab3369d43e1-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.053001 4956 scope.go:117] "RemoveContainer" containerID="814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.059402 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage15-crc" (UniqueName: "kubernetes.io/local-volume/local-storage15-crc") on node "crc" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.061250 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.064570 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.067643 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.075826 4956 scope.go:117] "RemoveContainer" containerID="5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.093004 4956 scope.go:117] "RemoveContainer" containerID="57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549" Nov 26 17:13:13 crc kubenswrapper[4956]: E1126 17:13:13.093735 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549\": container with ID starting with 57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549 not found: ID does not exist" containerID="57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.093783 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549"} err="failed to get container status \"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549\": rpc error: code = NotFound desc = could not find container \"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549\": container with ID starting with 57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549 not found: ID does not exist" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.093815 4956 scope.go:117] "RemoveContainer" containerID="814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d" Nov 26 17:13:13 crc kubenswrapper[4956]: E1126 17:13:13.094281 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d\": container with ID starting with 814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d not found: ID does not exist" containerID="814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.094323 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d"} err="failed to get container status \"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d\": rpc error: code = NotFound desc = could not find container \"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d\": container with ID starting with 814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d not found: ID does not exist" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.094346 4956 scope.go:117] "RemoveContainer" containerID="5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f" Nov 26 17:13:13 crc kubenswrapper[4956]: E1126 17:13:13.094652 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f\": container with ID starting with 5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f not found: ID does not exist" containerID="5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.094690 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f"} err="failed to get container status \"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f\": rpc error: code = NotFound desc = could not find container \"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f\": container with ID starting with 5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f not found: ID does not exist" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.094715 4956 scope.go:117] "RemoveContainer" containerID="57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.094961 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549"} err="failed to get container status \"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549\": rpc error: code = NotFound desc = could not find container \"57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549\": container with ID starting with 57e946edd33e837b93df9b9cf5c030104ef9c8036432bf5904ded3943b507549 not found: ID does not exist" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.094989 4956 scope.go:117] "RemoveContainer" containerID="814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.095245 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d"} err="failed to get container status \"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d\": rpc error: code = NotFound desc = could not find container \"814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d\": container with ID starting with 814b2262bad5aa6dd0b94dfa2599923433a4d1bda49dd86482e741cf07063a2d not found: ID does not exist" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.095282 4956 scope.go:117] "RemoveContainer" containerID="5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.095502 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f"} err="failed to get container status \"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f\": rpc error: code = NotFound desc = could not find container \"5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f\": container with ID starting with 5dcb773c92cf36a942d956846ebc00e9a380d0370cb973cfce5d163208cbb34f not found: ID does not exist" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.095529 4956 scope.go:117] "RemoveContainer" containerID="e6dcc44d126e5420818dd9c504b4375984f99cff37117259daa7588a10d187f7" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.122171 4956 scope.go:117] "RemoveContainer" containerID="7450db16bd9e52475d81036ff53fd1d1cba8bc9e378ed2be59ef9a8378e4c344" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.142107 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage15-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage15-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.142140 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.152178 4956 scope.go:117] "RemoveContainer" containerID="499706b812d257aa5178e3217a4b779fd63944df1f03c08859dd6080a135581c" Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.357856 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:13 crc kubenswrapper[4956]: I1126 17:13:13.365292 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:13 crc kubenswrapper[4956]: E1126 17:13:13.443517 4956 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod336b8230_0428_4ee1_bc0c_bab3369d43e1.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod336b8230_0428_4ee1_bc0c_bab3369d43e1.slice/crio-000e20dc49dcdfbd6a08be8365827fa5c53d2e91822e553aeda61de19b05b8cf\": RecentStats: unable to find data in memory cache]" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.597254 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pdnsb"] Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.602785 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-pdnsb"] Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.647272 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance15f5-account-delete-r2qhp"] Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.647961 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648044 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.648124 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648176 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.648303 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648364 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.648419 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648471 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.648535 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648591 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.648676 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648747 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.648830 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.648921 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649014 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649070 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649133 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649184 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649244 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649301 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649358 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649413 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649473 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649549 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649623 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649686 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649767 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.649837 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.649930 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650007 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.650063 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650115 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.650168 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650219 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: E1126 17:13:14.650280 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650340 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650642 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650726 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650790 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650849 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650922 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.650976 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651040 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651095 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651149 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b20c7ec5-2a9f-4a61-ad51-c09d5212a0a4" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651205 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7df26fee-12cb-4d53-a0d1-b33a106f141c" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651259 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651311 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651367 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651418 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651473 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651523 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="9cb132c4-0419-4520-ba16-e5f40377e318" containerName="glance-httpd" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651602 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" containerName="glance-api" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.651677 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ea8552e9-d8c9-4862-8d7b-9b019f8d4524" containerName="glance-log" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.652362 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.662421 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance15f5-account-delete-r2qhp"] Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.664786 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cs5t2\" (UniqueName: \"kubernetes.io/projected/5cd92e82-aef5-4533-ad98-6c64127302c1-kube-api-access-cs5t2\") pod \"glance15f5-account-delete-r2qhp\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.664886 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cd92e82-aef5-4533-ad98-6c64127302c1-operator-scripts\") pod \"glance15f5-account-delete-r2qhp\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.766456 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cs5t2\" (UniqueName: \"kubernetes.io/projected/5cd92e82-aef5-4533-ad98-6c64127302c1-kube-api-access-cs5t2\") pod \"glance15f5-account-delete-r2qhp\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.766915 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cd92e82-aef5-4533-ad98-6c64127302c1-operator-scripts\") pod \"glance15f5-account-delete-r2qhp\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.767843 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cd92e82-aef5-4533-ad98-6c64127302c1-operator-scripts\") pod \"glance15f5-account-delete-r2qhp\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:14 crc kubenswrapper[4956]: I1126 17:13:14.792084 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cs5t2\" (UniqueName: \"kubernetes.io/projected/5cd92e82-aef5-4533-ad98-6c64127302c1-kube-api-access-cs5t2\") pod \"glance15f5-account-delete-r2qhp\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:15 crc kubenswrapper[4956]: I1126 17:13:15.009450 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f8ade94-6d2b-4fa9-ab08-bf3ca268757c" path="/var/lib/kubelet/pods/1f8ade94-6d2b-4fa9-ab08-bf3ca268757c/volumes" Nov 26 17:13:15 crc kubenswrapper[4956]: I1126 17:13:15.010495 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:15 crc kubenswrapper[4956]: I1126 17:13:15.010639 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="336b8230-0428-4ee1-bc0c-bab3369d43e1" path="/var/lib/kubelet/pods/336b8230-0428-4ee1-bc0c-bab3369d43e1/volumes" Nov 26 17:13:15 crc kubenswrapper[4956]: I1126 17:13:15.011552 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4cbc07d2-febf-433f-87ed-4bcdff10b3f8" path="/var/lib/kubelet/pods/4cbc07d2-febf-433f-87ed-4bcdff10b3f8/volumes" Nov 26 17:13:15 crc kubenswrapper[4956]: I1126 17:13:15.424621 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance15f5-account-delete-r2qhp"] Nov 26 17:13:16 crc kubenswrapper[4956]: I1126 17:13:16.068357 4956 generic.go:334] "Generic (PLEG): container finished" podID="5cd92e82-aef5-4533-ad98-6c64127302c1" containerID="daca6791b8b1af0594e0b98f546c98e3dca5719bb46700a72aa7afd7b567cd8f" exitCode=0 Nov 26 17:13:16 crc kubenswrapper[4956]: I1126 17:13:16.068407 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" event={"ID":"5cd92e82-aef5-4533-ad98-6c64127302c1","Type":"ContainerDied","Data":"daca6791b8b1af0594e0b98f546c98e3dca5719bb46700a72aa7afd7b567cd8f"} Nov 26 17:13:16 crc kubenswrapper[4956]: I1126 17:13:16.068434 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" event={"ID":"5cd92e82-aef5-4533-ad98-6c64127302c1","Type":"ContainerStarted","Data":"bc011a78223719abfa82a8bf0e8ac6d5908e8f05cc26a0da998ba571ee084b91"} Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.403157 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.419876 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cd92e82-aef5-4533-ad98-6c64127302c1-operator-scripts\") pod \"5cd92e82-aef5-4533-ad98-6c64127302c1\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.420055 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cs5t2\" (UniqueName: \"kubernetes.io/projected/5cd92e82-aef5-4533-ad98-6c64127302c1-kube-api-access-cs5t2\") pod \"5cd92e82-aef5-4533-ad98-6c64127302c1\" (UID: \"5cd92e82-aef5-4533-ad98-6c64127302c1\") " Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.420774 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5cd92e82-aef5-4533-ad98-6c64127302c1-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "5cd92e82-aef5-4533-ad98-6c64127302c1" (UID: "5cd92e82-aef5-4533-ad98-6c64127302c1"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.421891 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/5cd92e82-aef5-4533-ad98-6c64127302c1-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.427106 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5cd92e82-aef5-4533-ad98-6c64127302c1-kube-api-access-cs5t2" (OuterVolumeSpecName: "kube-api-access-cs5t2") pod "5cd92e82-aef5-4533-ad98-6c64127302c1" (UID: "5cd92e82-aef5-4533-ad98-6c64127302c1"). InnerVolumeSpecName "kube-api-access-cs5t2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:17 crc kubenswrapper[4956]: I1126 17:13:17.523183 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cs5t2\" (UniqueName: \"kubernetes.io/projected/5cd92e82-aef5-4533-ad98-6c64127302c1-kube-api-access-cs5t2\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:18 crc kubenswrapper[4956]: I1126 17:13:18.083567 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" event={"ID":"5cd92e82-aef5-4533-ad98-6c64127302c1","Type":"ContainerDied","Data":"bc011a78223719abfa82a8bf0e8ac6d5908e8f05cc26a0da998ba571ee084b91"} Nov 26 17:13:18 crc kubenswrapper[4956]: I1126 17:13:18.083897 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bc011a78223719abfa82a8bf0e8ac6d5908e8f05cc26a0da998ba571ee084b91" Nov 26 17:13:18 crc kubenswrapper[4956]: I1126 17:13:18.083647 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance15f5-account-delete-r2qhp" Nov 26 17:13:19 crc kubenswrapper[4956]: I1126 17:13:19.670701 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-hhkv8"] Nov 26 17:13:19 crc kubenswrapper[4956]: I1126 17:13:19.677735 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-hhkv8"] Nov 26 17:13:19 crc kubenswrapper[4956]: I1126 17:13:19.694769 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-15f5-account-create-update-rkz5m"] Nov 26 17:13:19 crc kubenswrapper[4956]: I1126 17:13:19.701982 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance15f5-account-delete-r2qhp"] Nov 26 17:13:19 crc kubenswrapper[4956]: I1126 17:13:19.707605 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-15f5-account-create-update-rkz5m"] Nov 26 17:13:19 crc kubenswrapper[4956]: I1126 17:13:19.711486 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance15f5-account-delete-r2qhp"] Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.810700 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-2b8km"] Nov 26 17:13:20 crc kubenswrapper[4956]: E1126 17:13:20.811661 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5cd92e82-aef5-4533-ad98-6c64127302c1" containerName="mariadb-account-delete" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.811683 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5cd92e82-aef5-4533-ad98-6c64127302c1" containerName="mariadb-account-delete" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.811841 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5cd92e82-aef5-4533-ad98-6c64127302c1" containerName="mariadb-account-delete" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.812563 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.817912 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-363f-account-create-update-952hw"] Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.819558 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.825778 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.826656 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-363f-account-create-update-952hw"] Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.835826 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-2b8km"] Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.896962 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e50d91-c73e-45f0-8575-953e7621da5e-operator-scripts\") pod \"glance-363f-account-create-update-952hw\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.897038 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9ff972-55de-49b1-9ec0-38d333d97fca-operator-scripts\") pod \"glance-db-create-2b8km\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.897147 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wjbw9\" (UniqueName: \"kubernetes.io/projected/1f9ff972-55de-49b1-9ec0-38d333d97fca-kube-api-access-wjbw9\") pod \"glance-db-create-2b8km\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.897243 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xl4w7\" (UniqueName: \"kubernetes.io/projected/92e50d91-c73e-45f0-8575-953e7621da5e-kube-api-access-xl4w7\") pod \"glance-363f-account-create-update-952hw\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.999133 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e50d91-c73e-45f0-8575-953e7621da5e-operator-scripts\") pod \"glance-363f-account-create-update-952hw\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.999267 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9ff972-55de-49b1-9ec0-38d333d97fca-operator-scripts\") pod \"glance-db-create-2b8km\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.999378 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wjbw9\" (UniqueName: \"kubernetes.io/projected/1f9ff972-55de-49b1-9ec0-38d333d97fca-kube-api-access-wjbw9\") pod \"glance-db-create-2b8km\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:20 crc kubenswrapper[4956]: I1126 17:13:20.999515 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xl4w7\" (UniqueName: \"kubernetes.io/projected/92e50d91-c73e-45f0-8575-953e7621da5e-kube-api-access-xl4w7\") pod \"glance-363f-account-create-update-952hw\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:20.999975 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e50d91-c73e-45f0-8575-953e7621da5e-operator-scripts\") pod \"glance-363f-account-create-update-952hw\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.000018 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9ff972-55de-49b1-9ec0-38d333d97fca-operator-scripts\") pod \"glance-db-create-2b8km\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.004550 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1292cc2f-50be-4bdb-b9af-f939e15ca7d6" path="/var/lib/kubelet/pods/1292cc2f-50be-4bdb-b9af-f939e15ca7d6/volumes" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.005447 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5cd92e82-aef5-4533-ad98-6c64127302c1" path="/var/lib/kubelet/pods/5cd92e82-aef5-4533-ad98-6c64127302c1/volumes" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.006099 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1819f68-cff5-43a4-a73b-d37a47fd26e1" path="/var/lib/kubelet/pods/c1819f68-cff5-43a4-a73b-d37a47fd26e1/volumes" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.020301 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wjbw9\" (UniqueName: \"kubernetes.io/projected/1f9ff972-55de-49b1-9ec0-38d333d97fca-kube-api-access-wjbw9\") pod \"glance-db-create-2b8km\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.023378 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xl4w7\" (UniqueName: \"kubernetes.io/projected/92e50d91-c73e-45f0-8575-953e7621da5e-kube-api-access-xl4w7\") pod \"glance-363f-account-create-update-952hw\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.133189 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.153334 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.432853 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-363f-account-create-update-952hw"] Nov 26 17:13:21 crc kubenswrapper[4956]: I1126 17:13:21.555739 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-2b8km"] Nov 26 17:13:21 crc kubenswrapper[4956]: W1126 17:13:21.559479 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1f9ff972_55de_49b1_9ec0_38d333d97fca.slice/crio-b58078c8e31d4308897c070ef15e9b4e7072d51b0472dad7e4a3bbc3e5ee0509 WatchSource:0}: Error finding container b58078c8e31d4308897c070ef15e9b4e7072d51b0472dad7e4a3bbc3e5ee0509: Status 404 returned error can't find the container with id b58078c8e31d4308897c070ef15e9b4e7072d51b0472dad7e4a3bbc3e5ee0509 Nov 26 17:13:22 crc kubenswrapper[4956]: I1126 17:13:22.118632 4956 generic.go:334] "Generic (PLEG): container finished" podID="92e50d91-c73e-45f0-8575-953e7621da5e" containerID="aeea6d747ce568ce405549566998493dff831b6fbff9f50c80d92e7898d0a13c" exitCode=0 Nov 26 17:13:22 crc kubenswrapper[4956]: I1126 17:13:22.118698 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" event={"ID":"92e50d91-c73e-45f0-8575-953e7621da5e","Type":"ContainerDied","Data":"aeea6d747ce568ce405549566998493dff831b6fbff9f50c80d92e7898d0a13c"} Nov 26 17:13:22 crc kubenswrapper[4956]: I1126 17:13:22.118770 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" event={"ID":"92e50d91-c73e-45f0-8575-953e7621da5e","Type":"ContainerStarted","Data":"4f2964591eacd0eb481f8d3f03b79c6be1054409e9aa6f44b9b7ed49ce1c7c1d"} Nov 26 17:13:22 crc kubenswrapper[4956]: I1126 17:13:22.122489 4956 generic.go:334] "Generic (PLEG): container finished" podID="1f9ff972-55de-49b1-9ec0-38d333d97fca" containerID="b49fad21e2c286eb9f874fffb6845cd8208944ff5635e0a6e0e7fb4f4ec244eb" exitCode=0 Nov 26 17:13:22 crc kubenswrapper[4956]: I1126 17:13:22.122524 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-2b8km" event={"ID":"1f9ff972-55de-49b1-9ec0-38d333d97fca","Type":"ContainerDied","Data":"b49fad21e2c286eb9f874fffb6845cd8208944ff5635e0a6e0e7fb4f4ec244eb"} Nov 26 17:13:22 crc kubenswrapper[4956]: I1126 17:13:22.122563 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-2b8km" event={"ID":"1f9ff972-55de-49b1-9ec0-38d333d97fca","Type":"ContainerStarted","Data":"b58078c8e31d4308897c070ef15e9b4e7072d51b0472dad7e4a3bbc3e5ee0509"} Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.483847 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.489423 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.561044 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9ff972-55de-49b1-9ec0-38d333d97fca-operator-scripts\") pod \"1f9ff972-55de-49b1-9ec0-38d333d97fca\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.561178 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xl4w7\" (UniqueName: \"kubernetes.io/projected/92e50d91-c73e-45f0-8575-953e7621da5e-kube-api-access-xl4w7\") pod \"92e50d91-c73e-45f0-8575-953e7621da5e\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.561217 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wjbw9\" (UniqueName: \"kubernetes.io/projected/1f9ff972-55de-49b1-9ec0-38d333d97fca-kube-api-access-wjbw9\") pod \"1f9ff972-55de-49b1-9ec0-38d333d97fca\" (UID: \"1f9ff972-55de-49b1-9ec0-38d333d97fca\") " Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.561250 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e50d91-c73e-45f0-8575-953e7621da5e-operator-scripts\") pod \"92e50d91-c73e-45f0-8575-953e7621da5e\" (UID: \"92e50d91-c73e-45f0-8575-953e7621da5e\") " Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.561902 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/92e50d91-c73e-45f0-8575-953e7621da5e-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "92e50d91-c73e-45f0-8575-953e7621da5e" (UID: "92e50d91-c73e-45f0-8575-953e7621da5e"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.561925 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1f9ff972-55de-49b1-9ec0-38d333d97fca-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1f9ff972-55de-49b1-9ec0-38d333d97fca" (UID: "1f9ff972-55de-49b1-9ec0-38d333d97fca"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.562092 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/92e50d91-c73e-45f0-8575-953e7621da5e-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.562107 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1f9ff972-55de-49b1-9ec0-38d333d97fca-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.566416 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92e50d91-c73e-45f0-8575-953e7621da5e-kube-api-access-xl4w7" (OuterVolumeSpecName: "kube-api-access-xl4w7") pod "92e50d91-c73e-45f0-8575-953e7621da5e" (UID: "92e50d91-c73e-45f0-8575-953e7621da5e"). InnerVolumeSpecName "kube-api-access-xl4w7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.567641 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1f9ff972-55de-49b1-9ec0-38d333d97fca-kube-api-access-wjbw9" (OuterVolumeSpecName: "kube-api-access-wjbw9") pod "1f9ff972-55de-49b1-9ec0-38d333d97fca" (UID: "1f9ff972-55de-49b1-9ec0-38d333d97fca"). InnerVolumeSpecName "kube-api-access-wjbw9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.664100 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xl4w7\" (UniqueName: \"kubernetes.io/projected/92e50d91-c73e-45f0-8575-953e7621da5e-kube-api-access-xl4w7\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:23 crc kubenswrapper[4956]: I1126 17:13:23.664415 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wjbw9\" (UniqueName: \"kubernetes.io/projected/1f9ff972-55de-49b1-9ec0-38d333d97fca-kube-api-access-wjbw9\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:24 crc kubenswrapper[4956]: I1126 17:13:24.145054 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-2b8km" event={"ID":"1f9ff972-55de-49b1-9ec0-38d333d97fca","Type":"ContainerDied","Data":"b58078c8e31d4308897c070ef15e9b4e7072d51b0472dad7e4a3bbc3e5ee0509"} Nov 26 17:13:24 crc kubenswrapper[4956]: I1126 17:13:24.145128 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b58078c8e31d4308897c070ef15e9b4e7072d51b0472dad7e4a3bbc3e5ee0509" Nov 26 17:13:24 crc kubenswrapper[4956]: I1126 17:13:24.145235 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-2b8km" Nov 26 17:13:24 crc kubenswrapper[4956]: I1126 17:13:24.151096 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" event={"ID":"92e50d91-c73e-45f0-8575-953e7621da5e","Type":"ContainerDied","Data":"4f2964591eacd0eb481f8d3f03b79c6be1054409e9aa6f44b9b7ed49ce1c7c1d"} Nov 26 17:13:24 crc kubenswrapper[4956]: I1126 17:13:24.151156 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4f2964591eacd0eb481f8d3f03b79c6be1054409e9aa6f44b9b7ed49ce1c7c1d" Nov 26 17:13:24 crc kubenswrapper[4956]: I1126 17:13:24.151182 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-363f-account-create-update-952hw" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.041340 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-nb8kj"] Nov 26 17:13:26 crc kubenswrapper[4956]: E1126 17:13:26.041705 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92e50d91-c73e-45f0-8575-953e7621da5e" containerName="mariadb-account-create-update" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.041722 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="92e50d91-c73e-45f0-8575-953e7621da5e" containerName="mariadb-account-create-update" Nov 26 17:13:26 crc kubenswrapper[4956]: E1126 17:13:26.041751 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1f9ff972-55de-49b1-9ec0-38d333d97fca" containerName="mariadb-database-create" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.041760 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1f9ff972-55de-49b1-9ec0-38d333d97fca" containerName="mariadb-database-create" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.041950 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1f9ff972-55de-49b1-9ec0-38d333d97fca" containerName="mariadb-database-create" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.041971 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="92e50d91-c73e-45f0-8575-953e7621da5e" containerName="mariadb-account-create-update" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.042568 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.045325 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.055340 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7gfgv" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.058593 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nb8kj"] Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.103070 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-config-data\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.103906 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmpzt\" (UniqueName: \"kubernetes.io/projected/bd142fff-9783-4c61-b5ab-dab28086e6d4-kube-api-access-nmpzt\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.104116 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-db-sync-config-data\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.205803 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-config-data\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.205914 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmpzt\" (UniqueName: \"kubernetes.io/projected/bd142fff-9783-4c61-b5ab-dab28086e6d4-kube-api-access-nmpzt\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.205962 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-db-sync-config-data\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.211764 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-config-data\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.222504 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-db-sync-config-data\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.227338 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmpzt\" (UniqueName: \"kubernetes.io/projected/bd142fff-9783-4c61-b5ab-dab28086e6d4-kube-api-access-nmpzt\") pod \"glance-db-sync-nb8kj\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.362763 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:26 crc kubenswrapper[4956]: I1126 17:13:26.845184 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nb8kj"] Nov 26 17:13:27 crc kubenswrapper[4956]: I1126 17:13:27.176008 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nb8kj" event={"ID":"bd142fff-9783-4c61-b5ab-dab28086e6d4","Type":"ContainerStarted","Data":"737e1077fd85c595df30436bb5492b6721d98c21192a2567e75be19ebf3d6976"} Nov 26 17:13:28 crc kubenswrapper[4956]: I1126 17:13:28.187226 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nb8kj" event={"ID":"bd142fff-9783-4c61-b5ab-dab28086e6d4","Type":"ContainerStarted","Data":"18d72b5b430422cb8a3b35cc20b68088ca84c2df2e30d9e4428c2ce987544c09"} Nov 26 17:13:28 crc kubenswrapper[4956]: I1126 17:13:28.207100 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-nb8kj" podStartSLOduration=2.207070239 podStartE2EDuration="2.207070239s" podCreationTimestamp="2025-11-26 17:13:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:28.204828604 +0000 UTC m=+1253.900789156" watchObservedRunningTime="2025-11-26 17:13:28.207070239 +0000 UTC m=+1253.903030831" Nov 26 17:13:31 crc kubenswrapper[4956]: I1126 17:13:31.213361 4956 generic.go:334] "Generic (PLEG): container finished" podID="bd142fff-9783-4c61-b5ab-dab28086e6d4" containerID="18d72b5b430422cb8a3b35cc20b68088ca84c2df2e30d9e4428c2ce987544c09" exitCode=0 Nov 26 17:13:31 crc kubenswrapper[4956]: I1126 17:13:31.213451 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nb8kj" event={"ID":"bd142fff-9783-4c61-b5ab-dab28086e6d4","Type":"ContainerDied","Data":"18d72b5b430422cb8a3b35cc20b68088ca84c2df2e30d9e4428c2ce987544c09"} Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.534503 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.616275 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmpzt\" (UniqueName: \"kubernetes.io/projected/bd142fff-9783-4c61-b5ab-dab28086e6d4-kube-api-access-nmpzt\") pod \"bd142fff-9783-4c61-b5ab-dab28086e6d4\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.616343 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-config-data\") pod \"bd142fff-9783-4c61-b5ab-dab28086e6d4\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.616432 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-db-sync-config-data\") pod \"bd142fff-9783-4c61-b5ab-dab28086e6d4\" (UID: \"bd142fff-9783-4c61-b5ab-dab28086e6d4\") " Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.622465 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "bd142fff-9783-4c61-b5ab-dab28086e6d4" (UID: "bd142fff-9783-4c61-b5ab-dab28086e6d4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.623312 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd142fff-9783-4c61-b5ab-dab28086e6d4-kube-api-access-nmpzt" (OuterVolumeSpecName: "kube-api-access-nmpzt") pod "bd142fff-9783-4c61-b5ab-dab28086e6d4" (UID: "bd142fff-9783-4c61-b5ab-dab28086e6d4"). InnerVolumeSpecName "kube-api-access-nmpzt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.652719 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-config-data" (OuterVolumeSpecName: "config-data") pod "bd142fff-9783-4c61-b5ab-dab28086e6d4" (UID: "bd142fff-9783-4c61-b5ab-dab28086e6d4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.719138 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmpzt\" (UniqueName: \"kubernetes.io/projected/bd142fff-9783-4c61-b5ab-dab28086e6d4-kube-api-access-nmpzt\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.719203 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:32 crc kubenswrapper[4956]: I1126 17:13:32.719217 4956 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/bd142fff-9783-4c61-b5ab-dab28086e6d4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:33 crc kubenswrapper[4956]: I1126 17:13:33.230012 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-nb8kj" event={"ID":"bd142fff-9783-4c61-b5ab-dab28086e6d4","Type":"ContainerDied","Data":"737e1077fd85c595df30436bb5492b6721d98c21192a2567e75be19ebf3d6976"} Nov 26 17:13:33 crc kubenswrapper[4956]: I1126 17:13:33.230064 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-nb8kj" Nov 26 17:13:33 crc kubenswrapper[4956]: I1126 17:13:33.230069 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="737e1077fd85c595df30436bb5492b6721d98c21192a2567e75be19ebf3d6976" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.443927 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:13:34 crc kubenswrapper[4956]: E1126 17:13:34.444600 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd142fff-9783-4c61-b5ab-dab28086e6d4" containerName="glance-db-sync" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.444612 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd142fff-9783-4c61-b5ab-dab28086e6d4" containerName="glance-db-sync" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.444795 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd142fff-9783-4c61-b5ab-dab28086e6d4" containerName="glance-db-sync" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.445585 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.451446 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.451717 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-7gfgv" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.452005 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.468944 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.548950 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xb2pg\" (UniqueName: \"kubernetes.io/projected/be5457e4-228a-412f-af15-dfca3768af51-kube-api-access-xb2pg\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549017 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-sys\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549156 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549180 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-run\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549207 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549269 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549487 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-logs\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549513 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549539 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549577 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549604 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-scripts\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549634 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-config-data\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549668 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.549696 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-dev\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651075 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651150 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-dev\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651175 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xb2pg\" (UniqueName: \"kubernetes.io/projected/be5457e4-228a-412f-af15-dfca3768af51-kube-api-access-xb2pg\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651200 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-sys\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651270 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651292 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-run\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651314 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651382 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651401 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-logs\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651419 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651454 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651493 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651516 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-scripts\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.651550 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-config-data\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652275 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652326 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652428 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652501 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-dev\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652530 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-run\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652548 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-sys\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652575 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652771 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") device mount path \"/mnt/openstack/pv12\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652799 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.652970 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-logs\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.653025 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.658929 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-scripts\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.660356 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-config-data\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.677471 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xb2pg\" (UniqueName: \"kubernetes.io/projected/be5457e4-228a-412f-af15-dfca3768af51-kube-api-access-xb2pg\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.716324 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.719958 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-0\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.766721 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.978430 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.979916 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.984773 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 26 17:13:34 crc kubenswrapper[4956]: I1126 17:13:34.994383 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158548 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158631 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-logs\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158689 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158725 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-dev\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158751 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158776 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158805 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-sys\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158844 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ctwn9\" (UniqueName: \"kubernetes.io/projected/1ebabf9b-cf64-4f53-8cd5-e123135f7585-kube-api-access-ctwn9\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158893 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-run\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158924 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158946 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158966 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.158985 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.159017 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260681 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260761 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-dev\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260787 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260811 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260835 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-sys\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260862 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ctwn9\" (UniqueName: \"kubernetes.io/projected/1ebabf9b-cf64-4f53-8cd5-e123135f7585-kube-api-access-ctwn9\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260897 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-run\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260926 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260944 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260962 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260978 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.260996 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.261024 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.261046 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-logs\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.261528 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-logs\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262015 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-dev\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262429 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-run\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262086 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262168 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262480 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262307 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262380 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262376 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262249 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-sys\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.262412 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.268988 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-config-data\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.273349 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-scripts\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.288359 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ctwn9\" (UniqueName: \"kubernetes.io/projected/1ebabf9b-cf64-4f53-8cd5-e123135f7585-kube-api-access-ctwn9\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.293550 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.300183 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.301620 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.315539 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.579506 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:35 crc kubenswrapper[4956]: I1126 17:13:35.795704 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:35 crc kubenswrapper[4956]: W1126 17:13:35.807270 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1ebabf9b_cf64_4f53_8cd5_e123135f7585.slice/crio-906404094be2fd7ab2745ca6d22aa20f1ec05d45b57456493c31ce6f3571abd0 WatchSource:0}: Error finding container 906404094be2fd7ab2745ca6d22aa20f1ec05d45b57456493c31ce6f3571abd0: Status 404 returned error can't find the container with id 906404094be2fd7ab2745ca6d22aa20f1ec05d45b57456493c31ce6f3571abd0 Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.259045 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1ebabf9b-cf64-4f53-8cd5-e123135f7585","Type":"ContainerStarted","Data":"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66"} Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.259815 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1ebabf9b-cf64-4f53-8cd5-e123135f7585","Type":"ContainerStarted","Data":"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb"} Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.259832 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1ebabf9b-cf64-4f53-8cd5-e123135f7585","Type":"ContainerStarted","Data":"906404094be2fd7ab2745ca6d22aa20f1ec05d45b57456493c31ce6f3571abd0"} Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.260033 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-log" containerID="cri-o://cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb" gracePeriod=30 Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.260598 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-httpd" containerID="cri-o://da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66" gracePeriod=30 Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.263472 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"be5457e4-228a-412f-af15-dfca3768af51","Type":"ContainerStarted","Data":"810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6"} Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.263506 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"be5457e4-228a-412f-af15-dfca3768af51","Type":"ContainerStarted","Data":"984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3"} Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.263521 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"be5457e4-228a-412f-af15-dfca3768af51","Type":"ContainerStarted","Data":"51d34e2b9e1f3f15d81abd952ac25f781f136de62d19408e3ed6796aaadddb04"} Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.292052 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.292029088 podStartE2EDuration="3.292029088s" podCreationTimestamp="2025-11-26 17:13:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:36.28792628 +0000 UTC m=+1261.983886842" watchObservedRunningTime="2025-11-26 17:13:36.292029088 +0000 UTC m=+1261.987989640" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.326420 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.3264018379999998 podStartE2EDuration="2.326401838s" podCreationTimestamp="2025-11-26 17:13:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:36.320681713 +0000 UTC m=+1262.016642265" watchObservedRunningTime="2025-11-26 17:13:36.326401838 +0000 UTC m=+1262.022362390" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.645739 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788311 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-run\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788360 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-sys\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788418 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-run" (OuterVolumeSpecName: "run") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788428 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-lib-modules\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788453 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788480 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-nvme\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788530 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-scripts\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788563 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ctwn9\" (UniqueName: \"kubernetes.io/projected/1ebabf9b-cf64-4f53-8cd5-e123135f7585-kube-api-access-ctwn9\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788589 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-iscsi\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788628 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-dev\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788649 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-config-data\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788625 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-sys" (OuterVolumeSpecName: "sys") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788675 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-var-locks-brick\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788720 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788756 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-logs\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788783 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-httpd-run\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788799 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\" (UID: \"1ebabf9b-cf64-4f53-8cd5-e123135f7585\") " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.789108 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.789118 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.789126 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788685 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.788733 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.789805 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.790004 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-dev" (OuterVolumeSpecName: "dev") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.790254 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-logs" (OuterVolumeSpecName: "logs") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.790459 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.799694 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-scripts" (OuterVolumeSpecName: "scripts") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.805078 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.806092 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1ebabf9b-cf64-4f53-8cd5-e123135f7585-kube-api-access-ctwn9" (OuterVolumeSpecName: "kube-api-access-ctwn9") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "kube-api-access-ctwn9". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.806221 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.849141 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-config-data" (OuterVolumeSpecName: "config-data") pod "1ebabf9b-cf64-4f53-8cd5-e123135f7585" (UID: "1ebabf9b-cf64-4f53-8cd5-e123135f7585"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890390 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890440 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890452 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/1ebabf9b-cf64-4f53-8cd5-e123135f7585-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890468 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890477 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890485 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890497 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ctwn9\" (UniqueName: \"kubernetes.io/projected/1ebabf9b-cf64-4f53-8cd5-e123135f7585-kube-api-access-ctwn9\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890508 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890516 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890524 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1ebabf9b-cf64-4f53-8cd5-e123135f7585-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.890536 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/1ebabf9b-cf64-4f53-8cd5-e123135f7585-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.919091 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.927269 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 26 17:13:36 crc kubenswrapper[4956]: W1126 17:13:36.944213 4956 watcher.go:93] Error while processing event ("/sys/fs/cgroup/user.slice/user-0.slice/session-c32.scope": 0x40000100 == IN_CREATE|IN_ISDIR): inotify_add_watch /sys/fs/cgroup/user.slice/user-0.slice/session-c32.scope: no such file or directory Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.992041 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:36 crc kubenswrapper[4956]: I1126 17:13:36.992092 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.272474 4956 generic.go:334] "Generic (PLEG): container finished" podID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerID="da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66" exitCode=143 Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.272854 4956 generic.go:334] "Generic (PLEG): container finished" podID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerID="cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb" exitCode=143 Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.272591 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.272552 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1ebabf9b-cf64-4f53-8cd5-e123135f7585","Type":"ContainerDied","Data":"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66"} Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.273116 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1ebabf9b-cf64-4f53-8cd5-e123135f7585","Type":"ContainerDied","Data":"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb"} Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.273132 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"1ebabf9b-cf64-4f53-8cd5-e123135f7585","Type":"ContainerDied","Data":"906404094be2fd7ab2745ca6d22aa20f1ec05d45b57456493c31ce6f3571abd0"} Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.273153 4956 scope.go:117] "RemoveContainer" containerID="da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.303955 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.316599 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.318022 4956 scope.go:117] "RemoveContainer" containerID="cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.335533 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:37 crc kubenswrapper[4956]: E1126 17:13:37.335925 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-log" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.335965 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-log" Nov 26 17:13:37 crc kubenswrapper[4956]: E1126 17:13:37.336012 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-httpd" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.336021 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-httpd" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.336166 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-log" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.336194 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" containerName="glance-httpd" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.337078 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.341708 4956 scope.go:117] "RemoveContainer" containerID="da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.342334 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 26 17:13:37 crc kubenswrapper[4956]: E1126 17:13:37.342728 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66\": container with ID starting with da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66 not found: ID does not exist" containerID="da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.342782 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66"} err="failed to get container status \"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66\": rpc error: code = NotFound desc = could not find container \"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66\": container with ID starting with da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66 not found: ID does not exist" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.342828 4956 scope.go:117] "RemoveContainer" containerID="cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb" Nov 26 17:13:37 crc kubenswrapper[4956]: E1126 17:13:37.343424 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb\": container with ID starting with cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb not found: ID does not exist" containerID="cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.343474 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb"} err="failed to get container status \"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb\": rpc error: code = NotFound desc = could not find container \"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb\": container with ID starting with cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb not found: ID does not exist" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.343510 4956 scope.go:117] "RemoveContainer" containerID="da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.343917 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66"} err="failed to get container status \"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66\": rpc error: code = NotFound desc = could not find container \"da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66\": container with ID starting with da2363dd44b36ad892939cbda9abffdd2cd935186bdb495152ecbebb48e2bb66 not found: ID does not exist" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.343946 4956 scope.go:117] "RemoveContainer" containerID="cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.344190 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb"} err="failed to get container status \"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb\": rpc error: code = NotFound desc = could not find container \"cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb\": container with ID starting with cce4e36d77b9972ba4e5dbb136ce4e551cab99ee7908ec42b7015240650b18bb not found: ID does not exist" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.360666 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399140 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399221 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399250 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399303 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-run\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399406 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399443 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399480 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399783 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.399905 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cxjv7\" (UniqueName: \"kubernetes.io/projected/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-kube-api-access-cxjv7\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.400023 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-dev\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.400059 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-sys\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.400252 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-logs\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.400318 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.400341 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.501852 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-run\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.501936 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.501963 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.501996 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502047 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502074 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cxjv7\" (UniqueName: \"kubernetes.io/projected/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-kube-api-access-cxjv7\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502109 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-dev\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502127 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-sys\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502163 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-logs\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502183 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502205 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502232 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502263 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502282 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502559 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.502610 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-run\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.503080 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.503337 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.510183 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-dev\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.510298 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.510406 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") device mount path \"/mnt/openstack/pv20\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.510447 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.510523 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-sys\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.510810 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-logs\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.511114 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.525890 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-scripts\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.545239 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-config-data\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.549994 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cxjv7\" (UniqueName: \"kubernetes.io/projected/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-kube-api-access-cxjv7\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.567494 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.597358 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"glance-default-internal-api-0\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:37 crc kubenswrapper[4956]: I1126 17:13:37.655366 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:38 crc kubenswrapper[4956]: I1126 17:13:38.129556 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:13:38 crc kubenswrapper[4956]: I1126 17:13:38.281554 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a","Type":"ContainerStarted","Data":"d680bd8a75b4da9fa665b374b2e917812cfe3fd397a680dc89ece4f31ec9d356"} Nov 26 17:13:39 crc kubenswrapper[4956]: I1126 17:13:39.007644 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1ebabf9b-cf64-4f53-8cd5-e123135f7585" path="/var/lib/kubelet/pods/1ebabf9b-cf64-4f53-8cd5-e123135f7585/volumes" Nov 26 17:13:39 crc kubenswrapper[4956]: I1126 17:13:39.294403 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a","Type":"ContainerStarted","Data":"e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724"} Nov 26 17:13:39 crc kubenswrapper[4956]: I1126 17:13:39.294792 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a","Type":"ContainerStarted","Data":"03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39"} Nov 26 17:13:39 crc kubenswrapper[4956]: I1126 17:13:39.319711 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=2.319690282 podStartE2EDuration="2.319690282s" podCreationTimestamp="2025-11-26 17:13:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:39.314377739 +0000 UTC m=+1265.010338311" watchObservedRunningTime="2025-11-26 17:13:39.319690282 +0000 UTC m=+1265.015650834" Nov 26 17:13:44 crc kubenswrapper[4956]: I1126 17:13:44.767908 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:44 crc kubenswrapper[4956]: I1126 17:13:44.768860 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:44 crc kubenswrapper[4956]: I1126 17:13:44.798335 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:44 crc kubenswrapper[4956]: I1126 17:13:44.809126 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:45 crc kubenswrapper[4956]: I1126 17:13:45.339942 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:45 crc kubenswrapper[4956]: I1126 17:13:45.340002 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.647134 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.647771 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.651690 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.655916 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.655973 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.690429 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:47 crc kubenswrapper[4956]: I1126 17:13:47.709049 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:48 crc kubenswrapper[4956]: I1126 17:13:48.389943 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:48 crc kubenswrapper[4956]: I1126 17:13:48.389987 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:50 crc kubenswrapper[4956]: I1126 17:13:50.402471 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:13:50 crc kubenswrapper[4956]: I1126 17:13:50.403060 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:13:50 crc kubenswrapper[4956]: I1126 17:13:50.498694 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:50 crc kubenswrapper[4956]: I1126 17:13:50.504093 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.654628 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.657540 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.660662 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.662162 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.679042 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.686711 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782314 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-run\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782367 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-sys\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782397 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6lrvz\" (UniqueName: \"kubernetes.io/projected/7718c3f1-c165-41a2-a26d-d468915136b4-kube-api-access-6lrvz\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782428 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782523 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-run\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782562 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782761 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-logs\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782896 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-dev\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.782960 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783012 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783077 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783248 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-logs\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783462 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-dev\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783533 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783689 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783734 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783806 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783851 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.783929 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784025 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-scripts\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784073 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784105 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784242 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-scripts\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784316 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-sys\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784430 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-config-data\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784486 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdxkl\" (UniqueName: \"kubernetes.io/projected/cbfd374e-0a63-43e0-8a55-fe634b67b33f-kube-api-access-jdxkl\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784567 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-config-data\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.784624 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.798205 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.800162 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.803078 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.804570 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.819642 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886714 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-scripts\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886785 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qq2js\" (UniqueName: \"kubernetes.io/projected/d469cad6-31fb-4fab-b09a-6f4db845cbaa-kube-api-access-qq2js\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886820 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-sys\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886846 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-sys\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886888 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-config-data\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886908 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdxkl\" (UniqueName: \"kubernetes.io/projected/cbfd374e-0a63-43e0-8a55-fe634b67b33f-kube-api-access-jdxkl\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886906 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.886929 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-config-data\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887038 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-scripts\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887068 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887104 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887136 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887167 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6lrvz\" (UniqueName: \"kubernetes.io/projected/7718c3f1-c165-41a2-a26d-d468915136b4-kube-api-access-6lrvz\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887187 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-run\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887208 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-sys\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887236 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887279 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-config-data\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887333 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-run\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887415 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887439 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-logs\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887485 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-dev\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887506 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887532 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887555 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887583 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887631 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-logs\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887654 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887671 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-logs\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887705 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-sys\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887730 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887755 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-logs\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887942 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887976 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.887998 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrwnb\" (UniqueName: \"kubernetes.io/projected/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-kube-api-access-hrwnb\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888030 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-dev\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888049 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888074 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888107 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888144 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888182 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-run\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888203 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888219 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-config-data\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888233 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-run\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888255 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888287 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888305 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888325 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888343 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888361 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888388 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888391 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-sys\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888452 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888557 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-run\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.888571 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-scripts\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889060 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889104 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889104 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889119 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-logs\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889130 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-dev\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889191 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-scripts\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889219 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889247 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-nvme\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889298 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-lib-modules\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889329 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-iscsi\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889372 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-sys\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889393 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-dev\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889408 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-dev\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889451 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-var-locks-brick\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889142 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889667 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-run\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889728 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889770 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889784 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-logs\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.889790 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-dev\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.890077 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-httpd-run\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.890171 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.890242 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.890318 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.890396 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.896507 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-scripts\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.896788 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-scripts\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.900701 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-config-data\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.913352 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-config-data\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.918628 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6lrvz\" (UniqueName: \"kubernetes.io/projected/7718c3f1-c165-41a2-a26d-d468915136b4-kube-api-access-6lrvz\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.919741 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.919813 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdxkl\" (UniqueName: \"kubernetes.io/projected/cbfd374e-0a63-43e0-8a55-fe634b67b33f-kube-api-access-jdxkl\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.922458 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.929994 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-2\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.937168 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-external-api-1\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.990259 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992286 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992352 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992382 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrwnb\" (UniqueName: \"kubernetes.io/projected/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-kube-api-access-hrwnb\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992416 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992452 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992478 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992507 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-run\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992529 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-config-data\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992551 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-run\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992580 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992606 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992630 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992653 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-scripts\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992676 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-dev\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992703 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-dev\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992734 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992770 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qq2js\" (UniqueName: \"kubernetes.io/projected/d469cad6-31fb-4fab-b09a-6f4db845cbaa-kube-api-access-qq2js\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992800 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-sys\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992828 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-scripts\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992853 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992895 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992934 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-config-data\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.992970 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993000 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993022 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993042 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-logs\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993065 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-logs\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993088 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-sys\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993198 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-sys\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993250 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-lib-modules\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.993376 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994121 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-dev\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994268 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994388 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-run\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994500 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994780 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-httpd-run\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994780 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.994883 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-iscsi\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.995118 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-logs\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.995163 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.995169 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-sys\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.995201 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-var-locks-brick\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.995387 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.999238 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-scripts\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.999300 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-nvme\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:53 crc kubenswrapper[4956]: I1126 17:13:53.999393 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.000722 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-config-data\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.000842 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.000904 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-run\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.000939 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-dev\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.001463 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-logs\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.001527 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.003658 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-config-data\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.006391 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-scripts\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.007818 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.012204 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrwnb\" (UniqueName: \"kubernetes.io/projected/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-kube-api-access-hrwnb\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.012221 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qq2js\" (UniqueName: \"kubernetes.io/projected/d469cad6-31fb-4fab-b09a-6f4db845cbaa-kube-api-access-qq2js\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.032897 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.034591 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.044784 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-internal-api-1\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.088018 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-2\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.122925 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.130950 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.365554 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.437901 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"cbfd374e-0a63-43e0-8a55-fe634b67b33f","Type":"ContainerStarted","Data":"c289d9807f9c2c450ae8dcac8a8513b5a249cb0c65b01eadee04330e25c5877d"} Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.514284 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:13:54 crc kubenswrapper[4956]: W1126 17:13:54.519788 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7718c3f1_c165_41a2_a26d_d468915136b4.slice/crio-166f7438c0ab21a6138c12ae0c9118b6ee734b23d24d866a347fd1b2855b9303 WatchSource:0}: Error finding container 166f7438c0ab21a6138c12ae0c9118b6ee734b23d24d866a347fd1b2855b9303: Status 404 returned error can't find the container with id 166f7438c0ab21a6138c12ae0c9118b6ee734b23d24d866a347fd1b2855b9303 Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.680892 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:13:54 crc kubenswrapper[4956]: I1126 17:13:54.687277 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:13:54 crc kubenswrapper[4956]: W1126 17:13:54.693533 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod87549dcd_9fd7_4ca5_a226_8f2c2b80e6a1.slice/crio-7a793fad625aa3dadfa553e5bf329429096aabfb9fe5c8011a73856d55de9ded WatchSource:0}: Error finding container 7a793fad625aa3dadfa553e5bf329429096aabfb9fe5c8011a73856d55de9ded: Status 404 returned error can't find the container with id 7a793fad625aa3dadfa553e5bf329429096aabfb9fe5c8011a73856d55de9ded Nov 26 17:13:54 crc kubenswrapper[4956]: W1126 17:13:54.695157 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd469cad6_31fb_4fab_b09a_6f4db845cbaa.slice/crio-a1a97607e950093f111ed4876518b09889cab6dee1ac0d598433f6506d1d006d WatchSource:0}: Error finding container a1a97607e950093f111ed4876518b09889cab6dee1ac0d598433f6506d1d006d: Status 404 returned error can't find the container with id a1a97607e950093f111ed4876518b09889cab6dee1ac0d598433f6506d1d006d Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.450285 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"cbfd374e-0a63-43e0-8a55-fe634b67b33f","Type":"ContainerStarted","Data":"90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.450882 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"cbfd374e-0a63-43e0-8a55-fe634b67b33f","Type":"ContainerStarted","Data":"0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.452322 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d469cad6-31fb-4fab-b09a-6f4db845cbaa","Type":"ContainerStarted","Data":"1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.452351 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d469cad6-31fb-4fab-b09a-6f4db845cbaa","Type":"ContainerStarted","Data":"3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.452361 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d469cad6-31fb-4fab-b09a-6f4db845cbaa","Type":"ContainerStarted","Data":"a1a97607e950093f111ed4876518b09889cab6dee1ac0d598433f6506d1d006d"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.454826 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1","Type":"ContainerStarted","Data":"6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.454942 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1","Type":"ContainerStarted","Data":"29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.454960 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1","Type":"ContainerStarted","Data":"7a793fad625aa3dadfa553e5bf329429096aabfb9fe5c8011a73856d55de9ded"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.456810 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"7718c3f1-c165-41a2-a26d-d468915136b4","Type":"ContainerStarted","Data":"d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.456848 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"7718c3f1-c165-41a2-a26d-d468915136b4","Type":"ContainerStarted","Data":"6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.456859 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"7718c3f1-c165-41a2-a26d-d468915136b4","Type":"ContainerStarted","Data":"166f7438c0ab21a6138c12ae0c9118b6ee734b23d24d866a347fd1b2855b9303"} Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.483951 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=3.483923437 podStartE2EDuration="3.483923437s" podCreationTimestamp="2025-11-26 17:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:55.477153222 +0000 UTC m=+1281.173113794" watchObservedRunningTime="2025-11-26 17:13:55.483923437 +0000 UTC m=+1281.179883999" Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.510954 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-2" podStartSLOduration=3.510929205 podStartE2EDuration="3.510929205s" podCreationTimestamp="2025-11-26 17:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:55.503707417 +0000 UTC m=+1281.199667989" watchObservedRunningTime="2025-11-26 17:13:55.510929205 +0000 UTC m=+1281.206889767" Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.563104 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-2" podStartSLOduration=3.563079608 podStartE2EDuration="3.563079608s" podCreationTimestamp="2025-11-26 17:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:55.534406221 +0000 UTC m=+1281.230366773" watchObservedRunningTime="2025-11-26 17:13:55.563079608 +0000 UTC m=+1281.259040150" Nov 26 17:13:55 crc kubenswrapper[4956]: I1126 17:13:55.569325 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.569297357 podStartE2EDuration="3.569297357s" podCreationTimestamp="2025-11-26 17:13:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:13:55.561693188 +0000 UTC m=+1281.257653760" watchObservedRunningTime="2025-11-26 17:13:55.569297357 +0000 UTC m=+1281.265257909" Nov 26 17:14:03 crc kubenswrapper[4956]: I1126 17:14:03.991469 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:03 crc kubenswrapper[4956]: I1126 17:14:03.992031 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.008796 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.008889 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.025479 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.033049 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.034919 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.048604 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.123911 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.124266 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.132573 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.132604 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.150218 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.159157 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.170124 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.171943 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532490 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532533 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532546 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532555 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532563 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532573 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532585 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:04 crc kubenswrapper[4956]: I1126 17:14:04.532596 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.554032 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.554931 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.554255 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.555057 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.782523 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.782660 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.964421 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.964525 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.978608 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:06 crc kubenswrapper[4956]: I1126 17:14:06.981568 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:07 crc kubenswrapper[4956]: I1126 17:14:07.006564 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:07 crc kubenswrapper[4956]: I1126 17:14:07.107697 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:07 crc kubenswrapper[4956]: I1126 17:14:07.110182 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:07 crc kubenswrapper[4956]: I1126 17:14:07.500340 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.241288 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.256116 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.415176 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.423818 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.589058 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-log" containerID="cri-o://3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2" gracePeriod=30 Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.589137 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-httpd" containerID="cri-o://1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c" gracePeriod=30 Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.589154 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-log" containerID="cri-o://29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191" gracePeriod=30 Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.589268 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-httpd" containerID="cri-o://6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3" gracePeriod=30 Nov 26 17:14:08 crc kubenswrapper[4956]: I1126 17:14:08.597068 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-internal-api-2" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.120:9292/healthcheck\": EOF" Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.600196 4956 generic.go:334] "Generic (PLEG): container finished" podID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerID="3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2" exitCode=143 Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.600260 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d469cad6-31fb-4fab-b09a-6f4db845cbaa","Type":"ContainerDied","Data":"3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2"} Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.604124 4956 generic.go:334] "Generic (PLEG): container finished" podID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerID="29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191" exitCode=143 Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.604274 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1","Type":"ContainerDied","Data":"29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191"} Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.604393 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-log" containerID="cri-o://6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665" gracePeriod=30 Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.604448 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-2" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-httpd" containerID="cri-o://d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32" gracePeriod=30 Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.604814 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-log" containerID="cri-o://0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc" gracePeriod=30 Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.604928 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-httpd" containerID="cri-o://90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022" gracePeriod=30 Nov 26 17:14:09 crc kubenswrapper[4956]: I1126 17:14:09.613688 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.118:9292/healthcheck\": EOF" Nov 26 17:14:10 crc kubenswrapper[4956]: I1126 17:14:10.614557 4956 generic.go:334] "Generic (PLEG): container finished" podID="7718c3f1-c165-41a2-a26d-d468915136b4" containerID="6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665" exitCode=143 Nov 26 17:14:10 crc kubenswrapper[4956]: I1126 17:14:10.614604 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"7718c3f1-c165-41a2-a26d-d468915136b4","Type":"ContainerDied","Data":"6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665"} Nov 26 17:14:10 crc kubenswrapper[4956]: I1126 17:14:10.619130 4956 generic.go:334] "Generic (PLEG): container finished" podID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerID="0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc" exitCode=143 Nov 26 17:14:10 crc kubenswrapper[4956]: I1126 17:14:10.619203 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"cbfd374e-0a63-43e0-8a55-fe634b67b33f","Type":"ContainerDied","Data":"0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc"} Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.145764 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.283748 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.283819 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-httpd-run\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.283921 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-lib-modules\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.283944 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-dev\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.283978 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-iscsi\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.283997 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.284021 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-nvme\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.284137 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-dev" (OuterVolumeSpecName: "dev") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.284175 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.284137 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.284210 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.284383 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.285042 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-scripts\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.285127 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-logs\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.285161 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-run\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.285742 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-logs" (OuterVolumeSpecName: "logs") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.285912 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-run" (OuterVolumeSpecName: "run") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.288196 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-sys\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.288247 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-sys" (OuterVolumeSpecName: "sys") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.288329 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qq2js\" (UniqueName: \"kubernetes.io/projected/d469cad6-31fb-4fab-b09a-6f4db845cbaa-kube-api-access-qq2js\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.288381 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-var-locks-brick\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.288450 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-config-data\") pod \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\" (UID: \"d469cad6-31fb-4fab-b09a-6f4db845cbaa\") " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289365 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289382 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289391 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289400 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d469cad6-31fb-4fab-b09a-6f4db845cbaa-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289411 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289420 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289428 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.289438 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.290687 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.291212 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.293461 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-scripts" (OuterVolumeSpecName: "scripts") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.295236 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d469cad6-31fb-4fab-b09a-6f4db845cbaa-kube-api-access-qq2js" (OuterVolumeSpecName: "kube-api-access-qq2js") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "kube-api-access-qq2js". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.297090 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.339001 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-config-data" (OuterVolumeSpecName: "config-data") pod "d469cad6-31fb-4fab-b09a-6f4db845cbaa" (UID: "d469cad6-31fb-4fab-b09a-6f4db845cbaa"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.391638 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qq2js\" (UniqueName: \"kubernetes.io/projected/d469cad6-31fb-4fab-b09a-6f4db845cbaa-kube-api-access-qq2js\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.391679 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d469cad6-31fb-4fab-b09a-6f4db845cbaa-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.391692 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.391733 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.391747 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.391756 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d469cad6-31fb-4fab-b09a-6f4db845cbaa-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.407001 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.407367 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.493271 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.493309 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.655694 4956 generic.go:334] "Generic (PLEG): container finished" podID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerID="1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c" exitCode=0 Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.656161 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d469cad6-31fb-4fab-b09a-6f4db845cbaa","Type":"ContainerDied","Data":"1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c"} Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.656254 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"d469cad6-31fb-4fab-b09a-6f4db845cbaa","Type":"ContainerDied","Data":"a1a97607e950093f111ed4876518b09889cab6dee1ac0d598433f6506d1d006d"} Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.656337 4956 scope.go:117] "RemoveContainer" containerID="1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.656565 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.694122 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.703044 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.703665 4956 scope.go:117] "RemoveContainer" containerID="3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.724417 4956 scope.go:117] "RemoveContainer" containerID="1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c" Nov 26 17:14:12 crc kubenswrapper[4956]: E1126 17:14:12.724933 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c\": container with ID starting with 1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c not found: ID does not exist" containerID="1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.724978 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c"} err="failed to get container status \"1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c\": rpc error: code = NotFound desc = could not find container \"1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c\": container with ID starting with 1c242a5753cde937fc87cab59933be3813357bacbccd7c8d7f4dd848e17c6c4c not found: ID does not exist" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.725015 4956 scope.go:117] "RemoveContainer" containerID="3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2" Nov 26 17:14:12 crc kubenswrapper[4956]: E1126 17:14:12.726734 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2\": container with ID starting with 3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2 not found: ID does not exist" containerID="3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2" Nov 26 17:14:12 crc kubenswrapper[4956]: I1126 17:14:12.726766 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2"} err="failed to get container status \"3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2\": rpc error: code = NotFound desc = could not find container \"3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2\": container with ID starting with 3a0c07d7cd99508dce6db88ecc225d4972359628a6be33c015bdec7ca74426a2 not found: ID does not exist" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.005215 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" path="/var/lib/kubelet/pods/d469cad6-31fb-4fab-b09a-6f4db845cbaa/volumes" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.277602 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.420452 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-nvme\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.420938 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-logs\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.420958 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.420982 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-config-data\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421072 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-run\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421092 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-sys\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421165 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-lib-modules\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421181 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-httpd-run\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421215 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-scripts\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421247 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-var-locks-brick\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421288 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6lrvz\" (UniqueName: \"kubernetes.io/projected/7718c3f1-c165-41a2-a26d-d468915136b4-kube-api-access-6lrvz\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421310 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-dev\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421326 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-iscsi\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421350 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"7718c3f1-c165-41a2-a26d-d468915136b4\" (UID: \"7718c3f1-c165-41a2-a26d-d468915136b4\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421790 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.421879 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.422327 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-logs" (OuterVolumeSpecName: "logs") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.425852 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.425927 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-dev" (OuterVolumeSpecName: "dev") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.425884 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-run" (OuterVolumeSpecName: "run") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.425895 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-sys" (OuterVolumeSpecName: "sys") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.426274 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.426291 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.433453 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-scripts" (OuterVolumeSpecName: "scripts") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.433862 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.434001 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.439210 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7718c3f1-c165-41a2-a26d-d468915136b4-kube-api-access-6lrvz" (OuterVolumeSpecName: "kube-api-access-6lrvz") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "kube-api-access-6lrvz". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.463215 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.481830 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-config-data" (OuterVolumeSpecName: "config-data") pod "7718c3f1-c165-41a2-a26d-d468915136b4" (UID: "7718c3f1-c165-41a2-a26d-d468915136b4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.490819 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.522937 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6lrvz\" (UniqueName: \"kubernetes.io/projected/7718c3f1-c165-41a2-a26d-d468915136b4-kube-api-access-6lrvz\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.522978 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.522989 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523027 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523036 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523045 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523060 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523071 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523079 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523087 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523094 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523103 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7718c3f1-c165-41a2-a26d-d468915136b4-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523111 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7718c3f1-c165-41a2-a26d-d468915136b4-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.523118 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7718c3f1-c165-41a2-a26d-d468915136b4-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.539349 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.542900 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623757 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-dev\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623804 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623825 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-nvme\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623842 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-iscsi\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623900 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdxkl\" (UniqueName: \"kubernetes.io/projected/cbfd374e-0a63-43e0-8a55-fe634b67b33f-kube-api-access-jdxkl\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623933 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-var-locks-brick\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.623970 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-lib-modules\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624016 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-config-data\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624178 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-dev" (OuterVolumeSpecName: "dev") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624203 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624247 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624332 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624065 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624710 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrwnb\" (UniqueName: \"kubernetes.io/projected/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-kube-api-access-hrwnb\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624736 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-dev\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624761 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-httpd-run\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624782 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-iscsi\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624800 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-sys\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624818 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624827 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624844 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-dev" (OuterVolumeSpecName: "dev") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624855 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-logs\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624891 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.624942 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-sys" (OuterVolumeSpecName: "sys") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625002 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-lib-modules\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625065 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-logs\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625094 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-sys\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625115 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625154 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625196 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-nvme\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625237 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-scripts\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625240 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-sys" (OuterVolumeSpecName: "sys") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625268 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-var-locks-brick\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625302 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-config-data\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625299 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625346 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-httpd-run\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625381 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-run\") pod \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\" (UID: \"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625416 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-scripts\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625403 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625441 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-run\") pod \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\" (UID: \"cbfd374e-0a63-43e0-8a55-fe634b67b33f\") " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625471 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-run" (OuterVolumeSpecName: "run") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625486 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625594 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-logs" (OuterVolumeSpecName: "logs") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625645 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-run" (OuterVolumeSpecName: "run") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625933 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.625989 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-logs" (OuterVolumeSpecName: "logs") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626504 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626525 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626534 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626542 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626563 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626573 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626581 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626591 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626601 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626610 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626623 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626631 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626639 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626648 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626657 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626665 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626674 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/cbfd374e-0a63-43e0-8a55-fe634b67b33f-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.626682 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.629239 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.629262 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.629273 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cbfd374e-0a63-43e0-8a55-fe634b67b33f-kube-api-access-jdxkl" (OuterVolumeSpecName: "kube-api-access-jdxkl") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "kube-api-access-jdxkl". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.629308 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-kube-api-access-hrwnb" (OuterVolumeSpecName: "kube-api-access-hrwnb") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "kube-api-access-hrwnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.629350 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-scripts" (OuterVolumeSpecName: "scripts") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.631215 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-scripts" (OuterVolumeSpecName: "scripts") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.631312 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.631751 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance-cache") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.668965 4956 generic.go:334] "Generic (PLEG): container finished" podID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerID="6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3" exitCode=0 Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.669070 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1","Type":"ContainerDied","Data":"6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3"} Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.669112 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-2" event={"ID":"87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1","Type":"ContainerDied","Data":"7a793fad625aa3dadfa553e5bf329429096aabfb9fe5c8011a73856d55de9ded"} Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.669136 4956 scope.go:117] "RemoveContainer" containerID="6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.669276 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-2" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.669657 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-config-data" (OuterVolumeSpecName: "config-data") pod "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" (UID: "87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.675069 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-config-data" (OuterVolumeSpecName: "config-data") pod "cbfd374e-0a63-43e0-8a55-fe634b67b33f" (UID: "cbfd374e-0a63-43e0-8a55-fe634b67b33f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.676717 4956 generic.go:334] "Generic (PLEG): container finished" podID="7718c3f1-c165-41a2-a26d-d468915136b4" containerID="d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32" exitCode=0 Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.676847 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"7718c3f1-c165-41a2-a26d-d468915136b4","Type":"ContainerDied","Data":"d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32"} Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.676853 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-2" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.676898 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-2" event={"ID":"7718c3f1-c165-41a2-a26d-d468915136b4","Type":"ContainerDied","Data":"166f7438c0ab21a6138c12ae0c9118b6ee734b23d24d866a347fd1b2855b9303"} Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.681323 4956 generic.go:334] "Generic (PLEG): container finished" podID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerID="90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022" exitCode=0 Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.681378 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"cbfd374e-0a63-43e0-8a55-fe634b67b33f","Type":"ContainerDied","Data":"90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022"} Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.681396 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"cbfd374e-0a63-43e0-8a55-fe634b67b33f","Type":"ContainerDied","Data":"c289d9807f9c2c450ae8dcac8a8513b5a249cb0c65b01eadee04330e25c5877d"} Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.681410 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.696976 4956 scope.go:117] "RemoveContainer" containerID="29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.727356 4956 scope.go:117] "RemoveContainer" containerID="6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728430 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cbfd374e-0a63-43e0-8a55-fe634b67b33f-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728511 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728528 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728545 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728563 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728575 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728600 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728613 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdxkl\" (UniqueName: \"kubernetes.io/projected/cbfd374e-0a63-43e0-8a55-fe634b67b33f-kube-api-access-jdxkl\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728629 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cbfd374e-0a63-43e0-8a55-fe634b67b33f-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: E1126 17:14:13.728618 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3\": container with ID starting with 6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3 not found: ID does not exist" containerID="6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728670 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3"} err="failed to get container status \"6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3\": rpc error: code = NotFound desc = could not find container \"6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3\": container with ID starting with 6eb6b1b38d32a4477c19a14c85c04efa4029e138f5f990c3664336058afc5de3 not found: ID does not exist" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728696 4956 scope.go:117] "RemoveContainer" containerID="29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728645 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728765 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrwnb\" (UniqueName: \"kubernetes.io/projected/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1-kube-api-access-hrwnb\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.728802 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.742254 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.742319 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-2"] Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.747378 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 26 17:14:13 crc kubenswrapper[4956]: E1126 17:14:13.752469 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191\": container with ID starting with 29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191 not found: ID does not exist" containerID="29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.752562 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191"} err="failed to get container status \"29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191\": rpc error: code = NotFound desc = could not find container \"29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191\": container with ID starting with 29980a671b902388956fe8adb9601baefad81015632862d305a3fca116aad191 not found: ID does not exist" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.752626 4956 scope.go:117] "RemoveContainer" containerID="d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.753077 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.762470 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.768914 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.779237 4956 scope.go:117] "RemoveContainer" containerID="6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.786442 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.789937 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.804699 4956 scope.go:117] "RemoveContainer" containerID="d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32" Nov 26 17:14:13 crc kubenswrapper[4956]: E1126 17:14:13.805307 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32\": container with ID starting with d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32 not found: ID does not exist" containerID="d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.805346 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32"} err="failed to get container status \"d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32\": rpc error: code = NotFound desc = could not find container \"d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32\": container with ID starting with d8153962676047f1ef857d501e6857a77737c49bab576b1cc1dac331117e3e32 not found: ID does not exist" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.805375 4956 scope.go:117] "RemoveContainer" containerID="6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665" Nov 26 17:14:13 crc kubenswrapper[4956]: E1126 17:14:13.805936 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665\": container with ID starting with 6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665 not found: ID does not exist" containerID="6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.805963 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665"} err="failed to get container status \"6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665\": rpc error: code = NotFound desc = could not find container \"6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665\": container with ID starting with 6c14be3d3ee12d251343af547122916c1d71f04aef7f8cf2e129b6d78779f665 not found: ID does not exist" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.805978 4956 scope.go:117] "RemoveContainer" containerID="90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.839764 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.839819 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.839831 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.839844 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.842351 4956 scope.go:117] "RemoveContainer" containerID="0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.895964 4956 scope.go:117] "RemoveContainer" containerID="90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022" Nov 26 17:14:13 crc kubenswrapper[4956]: E1126 17:14:13.896415 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022\": container with ID starting with 90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022 not found: ID does not exist" containerID="90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.896455 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022"} err="failed to get container status \"90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022\": rpc error: code = NotFound desc = could not find container \"90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022\": container with ID starting with 90f2eec03816813378750981129249e4c179b7d6de293fc571973ff559de4022 not found: ID does not exist" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.896484 4956 scope.go:117] "RemoveContainer" containerID="0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc" Nov 26 17:14:13 crc kubenswrapper[4956]: E1126 17:14:13.897130 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc\": container with ID starting with 0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc not found: ID does not exist" containerID="0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc" Nov 26 17:14:13 crc kubenswrapper[4956]: I1126 17:14:13.897207 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc"} err="failed to get container status \"0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc\": rpc error: code = NotFound desc = could not find container \"0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc\": container with ID starting with 0a8c61ec194468f21bba114480e6816572231148fab7d6d4250192dfdd51e1dc not found: ID does not exist" Nov 26 17:14:14 crc kubenswrapper[4956]: I1126 17:14:14.007556 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:14:14 crc kubenswrapper[4956]: I1126 17:14:14.018212 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-2"] Nov 26 17:14:14 crc kubenswrapper[4956]: I1126 17:14:14.810085 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:14:14 crc kubenswrapper[4956]: I1126 17:14:14.811074 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-log" containerID="cri-o://03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39" gracePeriod=30 Nov 26 17:14:14 crc kubenswrapper[4956]: I1126 17:14:14.811184 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-httpd" containerID="cri-o://e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724" gracePeriod=30 Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.006686 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" path="/var/lib/kubelet/pods/7718c3f1-c165-41a2-a26d-d468915136b4/volumes" Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.007644 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" path="/var/lib/kubelet/pods/87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1/volumes" Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.008894 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" path="/var/lib/kubelet/pods/cbfd374e-0a63-43e0-8a55-fe634b67b33f/volumes" Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.275502 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.276770 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-log" containerID="cri-o://984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3" gracePeriod=30 Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.276892 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-httpd" containerID="cri-o://810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6" gracePeriod=30 Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.708251 4956 generic.go:334] "Generic (PLEG): container finished" podID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerID="03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39" exitCode=143 Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.708645 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a","Type":"ContainerDied","Data":"03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39"} Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.734466 4956 generic.go:334] "Generic (PLEG): container finished" podID="be5457e4-228a-412f-af15-dfca3768af51" containerID="984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3" exitCode=143 Nov 26 17:14:15 crc kubenswrapper[4956]: I1126 17:14:15.734532 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"be5457e4-228a-412f-af15-dfca3768af51","Type":"ContainerDied","Data":"984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3"} Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.363803 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.528683 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.528805 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-httpd-run\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.528831 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-sys\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.528939 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-config-data\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.528969 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-var-locks-brick\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529009 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-iscsi\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529023 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-sys" (OuterVolumeSpecName: "sys") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529103 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-scripts\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529160 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-nvme\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529228 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-lib-modules\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529107 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529127 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529268 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529292 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-run" (OuterVolumeSpecName: "run") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529257 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-run\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529356 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cxjv7\" (UniqueName: \"kubernetes.io/projected/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-kube-api-access-cxjv7\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529397 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-logs\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529417 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-dev\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529445 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") pod \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\" (UID: \"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529240 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529327 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529586 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-dev" (OuterVolumeSpecName: "dev") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529821 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-logs" (OuterVolumeSpecName: "logs") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529848 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529895 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529911 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529943 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529955 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529966 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529976 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.529986 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.537621 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage20-crc" (OuterVolumeSpecName: "glance") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "local-storage20-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.537649 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-scripts" (OuterVolumeSpecName: "scripts") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.540577 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance-cache") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.542397 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-kube-api-access-cxjv7" (OuterVolumeSpecName: "kube-api-access-cxjv7") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "kube-api-access-cxjv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.570572 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-config-data" (OuterVolumeSpecName: "config-data") pod "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" (UID: "4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.631961 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.632004 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.632014 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cxjv7\" (UniqueName: \"kubernetes.io/projected/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-kube-api-access-cxjv7\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.632025 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.632065 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.632078 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.653723 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage20-crc" (UniqueName: "kubernetes.io/local-volume/local-storage20-crc") on node "crc" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.655724 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.733377 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage20-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage20-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.733425 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.741519 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.765302 4956 generic.go:334] "Generic (PLEG): container finished" podID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerID="e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724" exitCode=0 Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.765398 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.765406 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a","Type":"ContainerDied","Data":"e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724"} Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.765468 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a","Type":"ContainerDied","Data":"d680bd8a75b4da9fa665b374b2e917812cfe3fd397a680dc89ece4f31ec9d356"} Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.765512 4956 scope.go:117] "RemoveContainer" containerID="e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.768834 4956 generic.go:334] "Generic (PLEG): container finished" podID="be5457e4-228a-412f-af15-dfca3768af51" containerID="810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6" exitCode=0 Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.768951 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"be5457e4-228a-412f-af15-dfca3768af51","Type":"ContainerDied","Data":"810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6"} Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.768991 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"be5457e4-228a-412f-af15-dfca3768af51","Type":"ContainerDied","Data":"51d34e2b9e1f3f15d81abd952ac25f781f136de62d19408e3ed6796aaadddb04"} Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.768999 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.811800 4956 scope.go:117] "RemoveContainer" containerID="03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.818016 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.825023 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.833059 4956 scope.go:117] "RemoveContainer" containerID="e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834149 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-httpd-run\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834197 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-config-data\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834232 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-nvme\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834302 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-dev\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834329 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-logs\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834349 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-run\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834372 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834394 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-sys\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834410 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-var-locks-brick\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834496 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834542 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-lib-modules\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834592 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-iscsi\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834636 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xb2pg\" (UniqueName: \"kubernetes.io/projected/be5457e4-228a-412f-af15-dfca3768af51-kube-api-access-xb2pg\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.834664 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-scripts\") pod \"be5457e4-228a-412f-af15-dfca3768af51\" (UID: \"be5457e4-228a-412f-af15-dfca3768af51\") " Nov 26 17:14:18 crc kubenswrapper[4956]: E1126 17:14:18.836295 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724\": container with ID starting with e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724 not found: ID does not exist" containerID="e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.836390 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724"} err="failed to get container status \"e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724\": rpc error: code = NotFound desc = could not find container \"e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724\": container with ID starting with e7c475351b21d2905b70c7492e2ca703a7f0d3398b27637301b203cf7eff8724 not found: ID does not exist" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.836437 4956 scope.go:117] "RemoveContainer" containerID="03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.836650 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-sys" (OuterVolumeSpecName: "sys") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.836783 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.837095 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.837136 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.837128 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.837604 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: E1126 17:14:18.837659 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39\": container with ID starting with 03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39 not found: ID does not exist" containerID="03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.837815 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39"} err="failed to get container status \"03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39\": rpc error: code = NotFound desc = could not find container \"03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39\": container with ID starting with 03dacbd5d05b2a488080b4629117bce314c3d199480d98a6b3fc7c16a82fec39 not found: ID does not exist" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.837839 4956 scope.go:117] "RemoveContainer" containerID="810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.838293 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-dev" (OuterVolumeSpecName: "dev") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.838336 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-run" (OuterVolumeSpecName: "run") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.838335 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-logs" (OuterVolumeSpecName: "logs") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.839203 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-scripts" (OuterVolumeSpecName: "scripts") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.839397 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage12-crc" (OuterVolumeSpecName: "glance") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "local-storage12-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.840193 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/be5457e4-228a-412f-af15-dfca3768af51-kube-api-access-xb2pg" (OuterVolumeSpecName: "kube-api-access-xb2pg") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "kube-api-access-xb2pg". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.840695 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance-cache") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.866227 4956 scope.go:117] "RemoveContainer" containerID="984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.871779 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-config-data" (OuterVolumeSpecName: "config-data") pod "be5457e4-228a-412f-af15-dfca3768af51" (UID: "be5457e4-228a-412f-af15-dfca3768af51"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.890702 4956 scope.go:117] "RemoveContainer" containerID="810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6" Nov 26 17:14:18 crc kubenswrapper[4956]: E1126 17:14:18.892006 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6\": container with ID starting with 810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6 not found: ID does not exist" containerID="810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.892051 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6"} err="failed to get container status \"810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6\": rpc error: code = NotFound desc = could not find container \"810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6\": container with ID starting with 810a34dcdfc777b30f701bc8ce0988e6bd3a9e51c19079052d8a22e6c6b53eb6 not found: ID does not exist" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.892082 4956 scope.go:117] "RemoveContainer" containerID="984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3" Nov 26 17:14:18 crc kubenswrapper[4956]: E1126 17:14:18.892369 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3\": container with ID starting with 984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3 not found: ID does not exist" containerID="984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.892402 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3"} err="failed to get container status \"984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3\": rpc error: code = NotFound desc = could not find container \"984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3\": container with ID starting with 984291c37778c2b39ca925e955e03e6893a325193bfb658c53388b891100afa3 not found: ID does not exist" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.935970 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936006 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xb2pg\" (UniqueName: \"kubernetes.io/projected/be5457e4-228a-412f-af15-dfca3768af51-kube-api-access-xb2pg\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936021 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936030 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936039 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/be5457e4-228a-412f-af15-dfca3768af51-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936047 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936057 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936066 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/be5457e4-228a-412f-af15-dfca3768af51-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936076 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936108 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936118 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936128 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936144 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.936153 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/be5457e4-228a-412f-af15-dfca3768af51-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.947997 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage12-crc" (UniqueName: "kubernetes.io/local-volume/local-storage12-crc") on node "crc" Nov 26 17:14:18 crc kubenswrapper[4956]: I1126 17:14:18.948019 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 26 17:14:19 crc kubenswrapper[4956]: I1126 17:14:19.008492 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" path="/var/lib/kubelet/pods/4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a/volumes" Nov 26 17:14:19 crc kubenswrapper[4956]: I1126 17:14:19.037532 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:19 crc kubenswrapper[4956]: I1126 17:14:19.037567 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:19 crc kubenswrapper[4956]: I1126 17:14:19.099013 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:19 crc kubenswrapper[4956]: I1126 17:14:19.106876 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.490858 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nb8kj"] Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.497223 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-nb8kj"] Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.525478 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance363f-account-delete-fgqhs"] Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.525961 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.525988 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526009 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526020 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526032 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526042 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526061 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526069 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526090 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526100 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526120 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526132 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526149 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526158 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526169 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526181 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526194 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526204 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526216 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526226 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526243 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526252 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: E1126 17:14:20.526264 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526273 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526541 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526566 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526581 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526595 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c2c1b3d-e6ce-4b3e-88d4-3d755fee643a" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526606 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526617 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="d469cad6-31fb-4fab-b09a-6f4db845cbaa" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526630 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="cbfd374e-0a63-43e0-8a55-fe634b67b33f" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526647 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526659 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="87549dcd-9fd7-4ca5-a226-8f2c2b80e6a1" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526702 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="be5457e4-228a-412f-af15-dfca3768af51" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526714 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-log" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.526729 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7718c3f1-c165-41a2-a26d-d468915136b4" containerName="glance-httpd" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.527592 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.544202 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance363f-account-delete-fgqhs"] Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.670307 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cbx4q\" (UniqueName: \"kubernetes.io/projected/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-kube-api-access-cbx4q\") pod \"glance363f-account-delete-fgqhs\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.670660 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-operator-scripts\") pod \"glance363f-account-delete-fgqhs\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.772685 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-operator-scripts\") pod \"glance363f-account-delete-fgqhs\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.772791 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cbx4q\" (UniqueName: \"kubernetes.io/projected/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-kube-api-access-cbx4q\") pod \"glance363f-account-delete-fgqhs\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.773760 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-operator-scripts\") pod \"glance363f-account-delete-fgqhs\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.804831 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cbx4q\" (UniqueName: \"kubernetes.io/projected/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-kube-api-access-cbx4q\") pod \"glance363f-account-delete-fgqhs\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:20 crc kubenswrapper[4956]: I1126 17:14:20.848228 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:21 crc kubenswrapper[4956]: I1126 17:14:21.004348 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd142fff-9783-4c61-b5ab-dab28086e6d4" path="/var/lib/kubelet/pods/bd142fff-9783-4c61-b5ab-dab28086e6d4/volumes" Nov 26 17:14:21 crc kubenswrapper[4956]: I1126 17:14:21.005538 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be5457e4-228a-412f-af15-dfca3768af51" path="/var/lib/kubelet/pods/be5457e4-228a-412f-af15-dfca3768af51/volumes" Nov 26 17:14:21 crc kubenswrapper[4956]: I1126 17:14:21.132334 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance363f-account-delete-fgqhs"] Nov 26 17:14:21 crc kubenswrapper[4956]: I1126 17:14:21.803239 4956 generic.go:334] "Generic (PLEG): container finished" podID="1294f3e4-673c-4c30-8b84-c7cb8d8512eb" containerID="7ad72aba18c790174f8236ebb4ed75f6dab833a1d5d7b3d3ab22b3829ed32d0e" exitCode=0 Nov 26 17:14:21 crc kubenswrapper[4956]: I1126 17:14:21.803451 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" event={"ID":"1294f3e4-673c-4c30-8b84-c7cb8d8512eb","Type":"ContainerDied","Data":"7ad72aba18c790174f8236ebb4ed75f6dab833a1d5d7b3d3ab22b3829ed32d0e"} Nov 26 17:14:21 crc kubenswrapper[4956]: I1126 17:14:21.803654 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" event={"ID":"1294f3e4-673c-4c30-8b84-c7cb8d8512eb","Type":"ContainerStarted","Data":"b06f3c6ffd0496669bfe2f28407cd56f2d57575af5e6100d18bc91b9d1423d01"} Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.147176 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.311174 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cbx4q\" (UniqueName: \"kubernetes.io/projected/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-kube-api-access-cbx4q\") pod \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.311577 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-operator-scripts\") pod \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\" (UID: \"1294f3e4-673c-4c30-8b84-c7cb8d8512eb\") " Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.312275 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "1294f3e4-673c-4c30-8b84-c7cb8d8512eb" (UID: "1294f3e4-673c-4c30-8b84-c7cb8d8512eb"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.312513 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.322563 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-kube-api-access-cbx4q" (OuterVolumeSpecName: "kube-api-access-cbx4q") pod "1294f3e4-673c-4c30-8b84-c7cb8d8512eb" (UID: "1294f3e4-673c-4c30-8b84-c7cb8d8512eb"). InnerVolumeSpecName "kube-api-access-cbx4q". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.414966 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cbx4q\" (UniqueName: \"kubernetes.io/projected/1294f3e4-673c-4c30-8b84-c7cb8d8512eb-kube-api-access-cbx4q\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.829804 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" event={"ID":"1294f3e4-673c-4c30-8b84-c7cb8d8512eb","Type":"ContainerDied","Data":"b06f3c6ffd0496669bfe2f28407cd56f2d57575af5e6100d18bc91b9d1423d01"} Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.829896 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b06f3c6ffd0496669bfe2f28407cd56f2d57575af5e6100d18bc91b9d1423d01" Nov 26 17:14:23 crc kubenswrapper[4956]: I1126 17:14:23.829978 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance363f-account-delete-fgqhs" Nov 26 17:14:25 crc kubenswrapper[4956]: I1126 17:14:25.557168 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-2b8km"] Nov 26 17:14:25 crc kubenswrapper[4956]: I1126 17:14:25.563959 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-2b8km"] Nov 26 17:14:25 crc kubenswrapper[4956]: I1126 17:14:25.580658 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-363f-account-create-update-952hw"] Nov 26 17:14:25 crc kubenswrapper[4956]: I1126 17:14:25.587484 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance363f-account-delete-fgqhs"] Nov 26 17:14:25 crc kubenswrapper[4956]: I1126 17:14:25.592367 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-363f-account-create-update-952hw"] Nov 26 17:14:25 crc kubenswrapper[4956]: I1126 17:14:25.596904 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance363f-account-delete-fgqhs"] Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.917394 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-4m8fj"] Nov 26 17:14:26 crc kubenswrapper[4956]: E1126 17:14:26.918675 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1294f3e4-673c-4c30-8b84-c7cb8d8512eb" containerName="mariadb-account-delete" Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.918710 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="1294f3e4-673c-4c30-8b84-c7cb8d8512eb" containerName="mariadb-account-delete" Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.919062 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="1294f3e4-673c-4c30-8b84-c7cb8d8512eb" containerName="mariadb-account-delete" Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.920221 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.927890 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-d650-account-create-update-72bds"] Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.929296 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.932352 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.935206 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-4m8fj"] Nov 26 17:14:26 crc kubenswrapper[4956]: I1126 17:14:26.946622 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d650-account-create-update-72bds"] Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.015566 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1294f3e4-673c-4c30-8b84-c7cb8d8512eb" path="/var/lib/kubelet/pods/1294f3e4-673c-4c30-8b84-c7cb8d8512eb/volumes" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.016294 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1f9ff972-55de-49b1-9ec0-38d333d97fca" path="/var/lib/kubelet/pods/1f9ff972-55de-49b1-9ec0-38d333d97fca/volumes" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.016788 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92e50d91-c73e-45f0-8575-953e7621da5e" path="/var/lib/kubelet/pods/92e50d91-c73e-45f0-8575-953e7621da5e/volumes" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.017379 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.018552 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.022054 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.022151 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-7fph6" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.022390 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.024279 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.031341 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.079699 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-operator-scripts\") pod \"glance-db-create-4m8fj\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.079776 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bg68w\" (UniqueName: \"kubernetes.io/projected/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-kube-api-access-bg68w\") pod \"glance-db-create-4m8fj\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.079808 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bqjj6\" (UniqueName: \"kubernetes.io/projected/c1be27b4-e03f-4b93-9f1e-2eb79b980791-kube-api-access-bqjj6\") pod \"glance-d650-account-create-update-72bds\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.080044 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1be27b4-e03f-4b93-9f1e-2eb79b980791-operator-scripts\") pod \"glance-d650-account-create-update-72bds\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182244 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bg68w\" (UniqueName: \"kubernetes.io/projected/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-kube-api-access-bg68w\") pod \"glance-db-create-4m8fj\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182346 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bqjj6\" (UniqueName: \"kubernetes.io/projected/c1be27b4-e03f-4b93-9f1e-2eb79b980791-kube-api-access-bqjj6\") pod \"glance-d650-account-create-update-72bds\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182420 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1be27b4-e03f-4b93-9f1e-2eb79b980791-operator-scripts\") pod \"glance-d650-account-create-update-72bds\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182498 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6z2jp\" (UniqueName: \"kubernetes.io/projected/d473f051-a35a-4c66-86ac-21c54348895e-kube-api-access-6z2jp\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182532 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config-secret\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182588 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-scripts\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182613 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.182733 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-operator-scripts\") pod \"glance-db-create-4m8fj\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.183525 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1be27b4-e03f-4b93-9f1e-2eb79b980791-operator-scripts\") pod \"glance-d650-account-create-update-72bds\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.183775 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-operator-scripts\") pod \"glance-db-create-4m8fj\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.207388 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bg68w\" (UniqueName: \"kubernetes.io/projected/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-kube-api-access-bg68w\") pod \"glance-db-create-4m8fj\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.208425 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bqjj6\" (UniqueName: \"kubernetes.io/projected/c1be27b4-e03f-4b93-9f1e-2eb79b980791-kube-api-access-bqjj6\") pod \"glance-d650-account-create-update-72bds\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.251497 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.276134 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.284527 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6z2jp\" (UniqueName: \"kubernetes.io/projected/d473f051-a35a-4c66-86ac-21c54348895e-kube-api-access-6z2jp\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.284616 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config-secret\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.284656 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-scripts\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.284685 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.286668 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.288601 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-scripts\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.291157 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config-secret\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.307060 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6z2jp\" (UniqueName: \"kubernetes.io/projected/d473f051-a35a-4c66-86ac-21c54348895e-kube-api-access-6z2jp\") pod \"openstackclient\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.341665 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.559919 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-d650-account-create-update-72bds"] Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.619244 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:14:27 crc kubenswrapper[4956]: W1126 17:14:27.625044 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd473f051_a35a_4c66_86ac_21c54348895e.slice/crio-c50b6e57c05a9d8c1e86688910820e795f35843add1cc754dd29631a5aa63a92 WatchSource:0}: Error finding container c50b6e57c05a9d8c1e86688910820e795f35843add1cc754dd29631a5aa63a92: Status 404 returned error can't find the container with id c50b6e57c05a9d8c1e86688910820e795f35843add1cc754dd29631a5aa63a92 Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.628045 4956 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.700907 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-4m8fj"] Nov 26 17:14:27 crc kubenswrapper[4956]: W1126 17:14:27.712351 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4fcfb1fa_bda6_4346_ae62_73f9df0c69dc.slice/crio-d4698e882b80963e5aa845b42b2c61e28051222e7253a92434ff7dac4946a197 WatchSource:0}: Error finding container d4698e882b80963e5aa845b42b2c61e28051222e7253a92434ff7dac4946a197: Status 404 returned error can't find the container with id d4698e882b80963e5aa845b42b2c61e28051222e7253a92434ff7dac4946a197 Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.885409 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4m8fj" event={"ID":"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc","Type":"ContainerStarted","Data":"0a27042a4bfa3e06491627e23c259bd6a35d0930c4424e580ce4f5f5c7358f93"} Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.885465 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4m8fj" event={"ID":"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc","Type":"ContainerStarted","Data":"d4698e882b80963e5aa845b42b2c61e28051222e7253a92434ff7dac4946a197"} Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.887205 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d473f051-a35a-4c66-86ac-21c54348895e","Type":"ContainerStarted","Data":"c50b6e57c05a9d8c1e86688910820e795f35843add1cc754dd29631a5aa63a92"} Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.890374 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" event={"ID":"c1be27b4-e03f-4b93-9f1e-2eb79b980791","Type":"ContainerStarted","Data":"bc5b585a847967f69417407326974c1faa7005b635172202cd81ea67765ccd00"} Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.890718 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" event={"ID":"c1be27b4-e03f-4b93-9f1e-2eb79b980791","Type":"ContainerStarted","Data":"076198ac94665c7ba4aebec5dc974998ae892bdde793250c7ce6e1ff59cb0cf2"} Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.907254 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-4m8fj" podStartSLOduration=1.907228662 podStartE2EDuration="1.907228662s" podCreationTimestamp="2025-11-26 17:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:27.902450323 +0000 UTC m=+1313.598410875" watchObservedRunningTime="2025-11-26 17:14:27.907228662 +0000 UTC m=+1313.603189214" Nov 26 17:14:27 crc kubenswrapper[4956]: I1126 17:14:27.919644 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" podStartSLOduration=1.919622562 podStartE2EDuration="1.919622562s" podCreationTimestamp="2025-11-26 17:14:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:27.918700725 +0000 UTC m=+1313.614661277" watchObservedRunningTime="2025-11-26 17:14:27.919622562 +0000 UTC m=+1313.615583114" Nov 26 17:14:28 crc kubenswrapper[4956]: I1126 17:14:28.914574 4956 generic.go:334] "Generic (PLEG): container finished" podID="c1be27b4-e03f-4b93-9f1e-2eb79b980791" containerID="bc5b585a847967f69417407326974c1faa7005b635172202cd81ea67765ccd00" exitCode=0 Nov 26 17:14:28 crc kubenswrapper[4956]: I1126 17:14:28.914641 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" event={"ID":"c1be27b4-e03f-4b93-9f1e-2eb79b980791","Type":"ContainerDied","Data":"bc5b585a847967f69417407326974c1faa7005b635172202cd81ea67765ccd00"} Nov 26 17:14:28 crc kubenswrapper[4956]: I1126 17:14:28.916908 4956 generic.go:334] "Generic (PLEG): container finished" podID="4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" containerID="0a27042a4bfa3e06491627e23c259bd6a35d0930c4424e580ce4f5f5c7358f93" exitCode=0 Nov 26 17:14:28 crc kubenswrapper[4956]: I1126 17:14:28.916957 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4m8fj" event={"ID":"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc","Type":"ContainerDied","Data":"0a27042a4bfa3e06491627e23c259bd6a35d0930c4424e580ce4f5f5c7358f93"} Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.356531 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.362931 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.442814 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1be27b4-e03f-4b93-9f1e-2eb79b980791-operator-scripts\") pod \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.442955 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bqjj6\" (UniqueName: \"kubernetes.io/projected/c1be27b4-e03f-4b93-9f1e-2eb79b980791-kube-api-access-bqjj6\") pod \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\" (UID: \"c1be27b4-e03f-4b93-9f1e-2eb79b980791\") " Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.443561 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c1be27b4-e03f-4b93-9f1e-2eb79b980791-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "c1be27b4-e03f-4b93-9f1e-2eb79b980791" (UID: "c1be27b4-e03f-4b93-9f1e-2eb79b980791"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.444266 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bg68w\" (UniqueName: \"kubernetes.io/projected/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-kube-api-access-bg68w\") pod \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.444308 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-operator-scripts\") pod \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\" (UID: \"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc\") " Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.444751 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/c1be27b4-e03f-4b93-9f1e-2eb79b980791-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.445717 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" (UID: "4fcfb1fa-bda6-4346-ae62-73f9df0c69dc"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.458253 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c1be27b4-e03f-4b93-9f1e-2eb79b980791-kube-api-access-bqjj6" (OuterVolumeSpecName: "kube-api-access-bqjj6") pod "c1be27b4-e03f-4b93-9f1e-2eb79b980791" (UID: "c1be27b4-e03f-4b93-9f1e-2eb79b980791"). InnerVolumeSpecName "kube-api-access-bqjj6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.458376 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-kube-api-access-bg68w" (OuterVolumeSpecName: "kube-api-access-bg68w") pod "4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" (UID: "4fcfb1fa-bda6-4346-ae62-73f9df0c69dc"). InnerVolumeSpecName "kube-api-access-bg68w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.547045 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bqjj6\" (UniqueName: \"kubernetes.io/projected/c1be27b4-e03f-4b93-9f1e-2eb79b980791-kube-api-access-bqjj6\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.547115 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bg68w\" (UniqueName: \"kubernetes.io/projected/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-kube-api-access-bg68w\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.547141 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.958560 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.959327 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-d650-account-create-update-72bds" event={"ID":"c1be27b4-e03f-4b93-9f1e-2eb79b980791","Type":"ContainerDied","Data":"076198ac94665c7ba4aebec5dc974998ae892bdde793250c7ce6e1ff59cb0cf2"} Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.959435 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="076198ac94665c7ba4aebec5dc974998ae892bdde793250c7ce6e1ff59cb0cf2" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.964266 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-4m8fj" event={"ID":"4fcfb1fa-bda6-4346-ae62-73f9df0c69dc","Type":"ContainerDied","Data":"d4698e882b80963e5aa845b42b2c61e28051222e7253a92434ff7dac4946a197"} Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.964298 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d4698e882b80963e5aa845b42b2c61e28051222e7253a92434ff7dac4946a197" Nov 26 17:14:31 crc kubenswrapper[4956]: I1126 17:14:30.964383 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-4m8fj" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.150315 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-x9hlb"] Nov 26 17:14:32 crc kubenswrapper[4956]: E1126 17:14:32.151317 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" containerName="mariadb-database-create" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.151338 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" containerName="mariadb-database-create" Nov 26 17:14:32 crc kubenswrapper[4956]: E1126 17:14:32.151358 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c1be27b4-e03f-4b93-9f1e-2eb79b980791" containerName="mariadb-account-create-update" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.151365 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="c1be27b4-e03f-4b93-9f1e-2eb79b980791" containerName="mariadb-account-create-update" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.151583 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" containerName="mariadb-database-create" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.151602 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="c1be27b4-e03f-4b93-9f1e-2eb79b980791" containerName="mariadb-account-create-update" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.152336 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.154624 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.155174 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-97vfs" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.169729 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x9hlb"] Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.291069 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-config-data\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.291152 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-db-sync-config-data\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.291224 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-97hnb\" (UniqueName: \"kubernetes.io/projected/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-kube-api-access-97hnb\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.392972 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-97hnb\" (UniqueName: \"kubernetes.io/projected/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-kube-api-access-97hnb\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.393109 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-config-data\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.393146 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-db-sync-config-data\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.404436 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-config-data\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.411574 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-db-sync-config-data\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.416006 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-97hnb\" (UniqueName: \"kubernetes.io/projected/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-kube-api-access-97hnb\") pod \"glance-db-sync-x9hlb\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:32 crc kubenswrapper[4956]: I1126 17:14:32.489234 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:36 crc kubenswrapper[4956]: I1126 17:14:36.461392 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x9hlb"] Nov 26 17:14:36 crc kubenswrapper[4956]: W1126 17:14:36.474198 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0037dfab_b9ac_4c2f_9f17_170f34a60b5a.slice/crio-754cf7b473283cc1b2fad15533b65c1188d7547fbf4c33d4df4d0b8a02ec3622 WatchSource:0}: Error finding container 754cf7b473283cc1b2fad15533b65c1188d7547fbf4c33d4df4d0b8a02ec3622: Status 404 returned error can't find the container with id 754cf7b473283cc1b2fad15533b65c1188d7547fbf4c33d4df4d0b8a02ec3622 Nov 26 17:14:37 crc kubenswrapper[4956]: I1126 17:14:37.031382 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d473f051-a35a-4c66-86ac-21c54348895e","Type":"ContainerStarted","Data":"2f2e7e1f79b9f23f1c20072ce2f9071c04f6bf83c98f34ffab3f58015fe36aa5"} Nov 26 17:14:37 crc kubenswrapper[4956]: I1126 17:14:37.039355 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x9hlb" event={"ID":"0037dfab-b9ac-4c2f-9f17-170f34a60b5a","Type":"ContainerStarted","Data":"754cf7b473283cc1b2fad15533b65c1188d7547fbf4c33d4df4d0b8a02ec3622"} Nov 26 17:14:37 crc kubenswrapper[4956]: I1126 17:14:37.061285 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.5690773780000002 podStartE2EDuration="11.061263056s" podCreationTimestamp="2025-11-26 17:14:26 +0000 UTC" firstStartedPulling="2025-11-26 17:14:27.627794482 +0000 UTC m=+1313.323755034" lastFinishedPulling="2025-11-26 17:14:36.11998017 +0000 UTC m=+1321.815940712" observedRunningTime="2025-11-26 17:14:37.060177114 +0000 UTC m=+1322.756137676" watchObservedRunningTime="2025-11-26 17:14:37.061263056 +0000 UTC m=+1322.757223608" Nov 26 17:14:38 crc kubenswrapper[4956]: I1126 17:14:38.050493 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x9hlb" event={"ID":"0037dfab-b9ac-4c2f-9f17-170f34a60b5a","Type":"ContainerStarted","Data":"913d2b9945bf65d004afee91480a370edf66ba6fd49aeb48ed80fbb0a58233d6"} Nov 26 17:14:38 crc kubenswrapper[4956]: I1126 17:14:38.072145 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-x9hlb" podStartSLOduration=6.072123916 podStartE2EDuration="6.072123916s" podCreationTimestamp="2025-11-26 17:14:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:38.06880799 +0000 UTC m=+1323.764768542" watchObservedRunningTime="2025-11-26 17:14:38.072123916 +0000 UTC m=+1323.768084468" Nov 26 17:14:39 crc kubenswrapper[4956]: I1126 17:14:39.550969 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:14:39 crc kubenswrapper[4956]: I1126 17:14:39.551407 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:14:41 crc kubenswrapper[4956]: I1126 17:14:41.083576 4956 generic.go:334] "Generic (PLEG): container finished" podID="0037dfab-b9ac-4c2f-9f17-170f34a60b5a" containerID="913d2b9945bf65d004afee91480a370edf66ba6fd49aeb48ed80fbb0a58233d6" exitCode=0 Nov 26 17:14:41 crc kubenswrapper[4956]: I1126 17:14:41.083657 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x9hlb" event={"ID":"0037dfab-b9ac-4c2f-9f17-170f34a60b5a","Type":"ContainerDied","Data":"913d2b9945bf65d004afee91480a370edf66ba6fd49aeb48ed80fbb0a58233d6"} Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.399664 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.468343 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-config-data\") pod \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.468393 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-db-sync-config-data\") pod \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.468505 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-97hnb\" (UniqueName: \"kubernetes.io/projected/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-kube-api-access-97hnb\") pod \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\" (UID: \"0037dfab-b9ac-4c2f-9f17-170f34a60b5a\") " Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.475202 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-kube-api-access-97hnb" (OuterVolumeSpecName: "kube-api-access-97hnb") pod "0037dfab-b9ac-4c2f-9f17-170f34a60b5a" (UID: "0037dfab-b9ac-4c2f-9f17-170f34a60b5a"). InnerVolumeSpecName "kube-api-access-97hnb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.475444 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "0037dfab-b9ac-4c2f-9f17-170f34a60b5a" (UID: "0037dfab-b9ac-4c2f-9f17-170f34a60b5a"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.506335 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-config-data" (OuterVolumeSpecName: "config-data") pod "0037dfab-b9ac-4c2f-9f17-170f34a60b5a" (UID: "0037dfab-b9ac-4c2f-9f17-170f34a60b5a"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.570954 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.571008 4956 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:42 crc kubenswrapper[4956]: I1126 17:14:42.571027 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-97hnb\" (UniqueName: \"kubernetes.io/projected/0037dfab-b9ac-4c2f-9f17-170f34a60b5a-kube-api-access-97hnb\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:43 crc kubenswrapper[4956]: I1126 17:14:43.105471 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-x9hlb" event={"ID":"0037dfab-b9ac-4c2f-9f17-170f34a60b5a","Type":"ContainerDied","Data":"754cf7b473283cc1b2fad15533b65c1188d7547fbf4c33d4df4d0b8a02ec3622"} Nov 26 17:14:43 crc kubenswrapper[4956]: I1126 17:14:43.105927 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="754cf7b473283cc1b2fad15533b65c1188d7547fbf4c33d4df4d0b8a02ec3622" Nov 26 17:14:43 crc kubenswrapper[4956]: I1126 17:14:43.105530 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-x9hlb" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.161797 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:14:44 crc kubenswrapper[4956]: E1126 17:14:44.162165 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0037dfab-b9ac-4c2f-9f17-170f34a60b5a" containerName="glance-db-sync" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.162180 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0037dfab-b9ac-4c2f-9f17-170f34a60b5a" containerName="glance-db-sync" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.162327 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="0037dfab-b9ac-4c2f-9f17-170f34a60b5a" containerName="glance-db-sync" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.163235 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.167732 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-97vfs" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.168176 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.168428 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-external-config-data" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.236104 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.297263 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-config-data\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.297328 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.297358 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.297391 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.297761 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-scripts\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.297921 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-sys\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298032 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298150 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298187 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-run\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298296 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-dev\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298327 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-logs\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298376 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cssg6\" (UniqueName: \"kubernetes.io/projected/e164968b-dfde-4d68-b73e-632722876b76-kube-api-access-cssg6\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298454 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.298635 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400670 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400730 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400770 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400793 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-scripts\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400813 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-sys\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400839 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400885 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-run\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400896 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-var-locks-brick\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400903 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.400966 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-dev\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401000 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-logs\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401021 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cssg6\" (UniqueName: \"kubernetes.io/projected/e164968b-dfde-4d68-b73e-632722876b76-kube-api-access-cssg6\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401038 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401079 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401107 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-config-data\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401314 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") device mount path \"/mnt/openstack/pv10\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401318 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-httpd-run\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401318 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") device mount path \"/mnt/openstack/pv11\"" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401652 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-iscsi\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401706 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-sys\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401652 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-lib-modules\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401752 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-run\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401749 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-dev\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.401785 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-nvme\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.402248 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-logs\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.406918 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-scripts\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.412072 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-config-data\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.421261 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cssg6\" (UniqueName: \"kubernetes.io/projected/e164968b-dfde-4d68-b73e-632722876b76-kube-api-access-cssg6\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.424003 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.436904 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"glance-default-external-api-1\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.458812 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.460330 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.463783 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-internal-config-data" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.484838 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.490371 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502193 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-sys\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502234 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-dev\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502262 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502305 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502345 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502369 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502385 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502402 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2pvp7\" (UniqueName: \"kubernetes.io/projected/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-kube-api-access-2pvp7\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502429 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502450 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502471 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502493 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502512 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-run\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.502551 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.517392 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.563121 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.586197 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.604724 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.604798 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.604819 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2pvp7\" (UniqueName: \"kubernetes.io/projected/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-kube-api-access-2pvp7\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.604948 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.604975 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.604996 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605042 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605064 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-run\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605102 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605143 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-sys\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605189 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-dev\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605214 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605288 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.605348 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.606509 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-logs\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.607011 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.607269 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.607380 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.607433 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.607718 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.608476 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-dev\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.608559 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-run\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.608598 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.608632 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-sys\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.609449 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.611186 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.612627 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.612762 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.617418 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.623942 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.632721 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2pvp7\" (UniqueName: \"kubernetes.io/projected/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-kube-api-access-2pvp7\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.639968 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.661073 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.706968 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-dev\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.707023 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.707051 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.707080 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-config-data\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.707101 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709052 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709142 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709182 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-sys\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709233 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fcl7s\" (UniqueName: \"kubernetes.io/projected/5a647b90-78f6-42fe-a636-493b9ead83b9-kube-api-access-fcl7s\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709539 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709589 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709687 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-logs\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.709773 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-run\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.710000 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-scripts\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.811850 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.811922 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.811941 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-sys\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.811963 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.811991 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812019 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812044 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-logs\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812084 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812259 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812320 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-logs\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812414 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-run\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812477 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-scripts\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812527 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812585 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812641 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-dev\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812672 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2m5w\" (UniqueName: \"kubernetes.io/projected/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-kube-api-access-l2m5w\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812698 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812735 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812776 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812800 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812838 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812886 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-config-data\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812915 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812969 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-dev\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.812996 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813018 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-run\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813044 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813079 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-sys\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813104 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-run\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813169 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813177 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fcl7s\" (UniqueName: \"kubernetes.io/projected/5a647b90-78f6-42fe-a636-493b9ead83b9-kube-api-access-fcl7s\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813235 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813346 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813365 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-logs\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813369 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813338 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-sys\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813435 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.813672 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.814303 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.819006 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-scripts\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.826563 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-config-data\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.832105 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-dev\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.838744 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.841511 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fcl7s\" (UniqueName: \"kubernetes.io/projected/5a647b90-78f6-42fe-a636-493b9ead83b9-kube-api-access-fcl7s\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.851694 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.914920 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915373 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915422 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2m5w\" (UniqueName: \"kubernetes.io/projected/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-kube-api-access-l2m5w\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915463 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915482 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915512 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915552 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-dev\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915583 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-run\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915634 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915651 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915667 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-sys\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915689 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915714 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.915734 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-logs\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.916267 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-logs\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.922045 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-dev\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.922152 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.922221 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.922327 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-run\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.922449 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.923455 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.923766 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-scripts\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.923850 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.923919 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-sys\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.925098 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.927244 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.930134 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-config-data\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.944135 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.948812 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2m5w\" (UniqueName: \"kubernetes.io/projected/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-kube-api-access-l2m5w\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.955923 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:44 crc kubenswrapper[4956]: I1126 17:14:44.959758 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.093473 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.125116 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e164968b-dfde-4d68-b73e-632722876b76","Type":"ContainerStarted","Data":"92fbaaf108139cb95a45215dbf12639f90a8e4951b69d89701ed22f900013493"} Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.249246 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.295729 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.307613 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.315672 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:14:45 crc kubenswrapper[4956]: I1126 17:14:45.762081 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:45 crc kubenswrapper[4956]: W1126 17:14:45.775613 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4c38ee74_c6a3_4ea1_9285_bc85ec9f1b3c.slice/crio-dc016d71d7127772cd969e77fb5258bbf832ad9cf9fd2991c949d5e898327ced WatchSource:0}: Error finding container dc016d71d7127772cd969e77fb5258bbf832ad9cf9fd2991c949d5e898327ced: Status 404 returned error can't find the container with id dc016d71d7127772cd969e77fb5258bbf832ad9cf9fd2991c949d5e898327ced Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.134447 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e164968b-dfde-4d68-b73e-632722876b76","Type":"ContainerStarted","Data":"18551c8bd2a6a38f12abf66c667eb04138078566eef08a87c1dc50ecb7afa55d"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.135331 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e164968b-dfde-4d68-b73e-632722876b76","Type":"ContainerStarted","Data":"5767b20b397a28d1c45dcf61e4de7c5a4a7f897fce042c2a43444107bb4e6e59"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.139512 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c","Type":"ContainerStarted","Data":"40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.139576 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c","Type":"ContainerStarted","Data":"78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.139588 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c","Type":"ContainerStarted","Data":"dc016d71d7127772cd969e77fb5258bbf832ad9cf9fd2991c949d5e898327ced"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.143037 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8","Type":"ContainerStarted","Data":"a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.143099 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8","Type":"ContainerStarted","Data":"da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.143116 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8","Type":"ContainerStarted","Data":"201e3b2cfff8b4f2cb63208e56e0e55a0371354103bec5152762fc678a97bf79"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.146859 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"5a647b90-78f6-42fe-a636-493b9ead83b9","Type":"ContainerStarted","Data":"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.146923 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"5a647b90-78f6-42fe-a636-493b9ead83b9","Type":"ContainerStarted","Data":"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.146941 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"5a647b90-78f6-42fe-a636-493b9ead83b9","Type":"ContainerStarted","Data":"4104025bd2f8f76f51523f152a39ee94be89db7ce7c88d2e3a217e6a9985f816"} Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.147010 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-log" containerID="cri-o://dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e" gracePeriod=30 Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.147047 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-httpd" containerID="cri-o://e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a" gracePeriod=30 Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.171276 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-1" podStartSLOduration=2.171247568 podStartE2EDuration="2.171247568s" podCreationTimestamp="2025-11-26 17:14:44 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:46.167441648 +0000 UTC m=+1331.863402220" watchObservedRunningTime="2025-11-26 17:14:46.171247568 +0000 UTC m=+1331.867208120" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.206801 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=3.206780452 podStartE2EDuration="3.206780452s" podCreationTimestamp="2025-11-26 17:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:46.202070895 +0000 UTC m=+1331.898031447" watchObservedRunningTime="2025-11-26 17:14:46.206780452 +0000 UTC m=+1331.902741004" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.237100 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=3.237069954 podStartE2EDuration="3.237069954s" podCreationTimestamp="2025-11-26 17:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:46.22834874 +0000 UTC m=+1331.924309312" watchObservedRunningTime="2025-11-26 17:14:46.237069954 +0000 UTC m=+1331.933030526" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.257930 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=3.25790426 podStartE2EDuration="3.25790426s" podCreationTimestamp="2025-11-26 17:14:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:46.255234442 +0000 UTC m=+1331.951195004" watchObservedRunningTime="2025-11-26 17:14:46.25790426 +0000 UTC m=+1331.953864812" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.587545 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.678887 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-nvme\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679003 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-logs\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679031 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-var-locks-brick\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679033 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679062 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-sys\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679093 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-iscsi\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679102 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679137 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-sys" (OuterVolumeSpecName: "sys") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679158 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679195 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-scripts\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679238 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-httpd-run\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679265 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-lib-modules\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679304 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679343 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-run\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679416 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-run" (OuterVolumeSpecName: "run") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679445 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-config-data\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679447 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-logs" (OuterVolumeSpecName: "logs") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679513 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679583 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-dev\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679627 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcl7s\" (UniqueName: \"kubernetes.io/projected/5a647b90-78f6-42fe-a636-493b9ead83b9-kube-api-access-fcl7s\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679661 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"5a647b90-78f6-42fe-a636-493b9ead83b9\" (UID: \"5a647b90-78f6-42fe-a636-493b9ead83b9\") " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679778 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.679835 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-dev" (OuterVolumeSpecName: "dev") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680536 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680552 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680561 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680571 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680581 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680593 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5a647b90-78f6-42fe-a636-493b9ead83b9-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680602 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680612 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.680625 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/5a647b90-78f6-42fe-a636-493b9ead83b9-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.686477 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-scripts" (OuterVolumeSpecName: "scripts") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.686493 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5a647b90-78f6-42fe-a636-493b9ead83b9-kube-api-access-fcl7s" (OuterVolumeSpecName: "kube-api-access-fcl7s") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "kube-api-access-fcl7s". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.690096 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.690846 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.747025 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-config-data" (OuterVolumeSpecName: "config-data") pod "5a647b90-78f6-42fe-a636-493b9ead83b9" (UID: "5a647b90-78f6-42fe-a636-493b9ead83b9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.783067 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.783184 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.783241 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5a647b90-78f6-42fe-a636-493b9ead83b9-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.783292 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcl7s\" (UniqueName: \"kubernetes.io/projected/5a647b90-78f6-42fe-a636-493b9ead83b9-kube-api-access-fcl7s\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.783352 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.798387 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.799520 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.885648 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:46 crc kubenswrapper[4956]: I1126 17:14:46.886069 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159168 4956 generic.go:334] "Generic (PLEG): container finished" podID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerID="e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a" exitCode=143 Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159244 4956 generic.go:334] "Generic (PLEG): container finished" podID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerID="dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e" exitCode=143 Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159315 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"5a647b90-78f6-42fe-a636-493b9ead83b9","Type":"ContainerDied","Data":"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a"} Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159378 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159436 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"5a647b90-78f6-42fe-a636-493b9ead83b9","Type":"ContainerDied","Data":"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e"} Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159466 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"5a647b90-78f6-42fe-a636-493b9ead83b9","Type":"ContainerDied","Data":"4104025bd2f8f76f51523f152a39ee94be89db7ce7c88d2e3a217e6a9985f816"} Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.159504 4956 scope.go:117] "RemoveContainer" containerID="e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.185858 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.197831 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.206203 4956 scope.go:117] "RemoveContainer" containerID="dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.228315 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:47 crc kubenswrapper[4956]: E1126 17:14:47.229596 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-httpd" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.229657 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-httpd" Nov 26 17:14:47 crc kubenswrapper[4956]: E1126 17:14:47.229680 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-log" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.229687 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-log" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.230034 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-httpd" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.230058 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" containerName="glance-log" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.231531 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.237523 4956 scope.go:117] "RemoveContainer" containerID="e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a" Nov 26 17:14:47 crc kubenswrapper[4956]: E1126 17:14:47.244357 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a\": container with ID starting with e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a not found: ID does not exist" containerID="e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.244431 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a"} err="failed to get container status \"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a\": rpc error: code = NotFound desc = could not find container \"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a\": container with ID starting with e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a not found: ID does not exist" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.244468 4956 scope.go:117] "RemoveContainer" containerID="dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e" Nov 26 17:14:47 crc kubenswrapper[4956]: E1126 17:14:47.245413 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e\": container with ID starting with dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e not found: ID does not exist" containerID="dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.245542 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e"} err="failed to get container status \"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e\": rpc error: code = NotFound desc = could not find container \"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e\": container with ID starting with dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e not found: ID does not exist" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.245698 4956 scope.go:117] "RemoveContainer" containerID="e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.246284 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a"} err="failed to get container status \"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a\": rpc error: code = NotFound desc = could not find container \"e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a\": container with ID starting with e275c36aa575f53084ddceddaed85616b9b5863cdb4e763266179ec409e4853a not found: ID does not exist" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.246310 4956 scope.go:117] "RemoveContainer" containerID="dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.246725 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e"} err="failed to get container status \"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e\": rpc error: code = NotFound desc = could not find container \"dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e\": container with ID starting with dd5adee69b4a8fe1e0ad5876b9f2e5329ddfc9423a91c4807d98066f0a37a93e not found: ID does not exist" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.253355 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396075 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-scripts\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396506 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396532 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-logs\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396555 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-run\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396577 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396602 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396624 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396645 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-config-data\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396665 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396681 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396714 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-dev\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396732 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scxgw\" (UniqueName: \"kubernetes.io/projected/6e2d74c2-ac47-4731-a3e5-641274858dad-kube-api-access-scxgw\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396764 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.396799 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-sys\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498049 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-dev\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498103 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scxgw\" (UniqueName: \"kubernetes.io/projected/6e2d74c2-ac47-4731-a3e5-641274858dad-kube-api-access-scxgw\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498150 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498195 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-sys\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498186 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-dev\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498224 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-scripts\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498274 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-sys\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498277 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-nvme\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498391 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498462 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-run\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498514 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-logs\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498550 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498598 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498624 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498672 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498692 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-config-data\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.498719 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499021 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499234 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-logs\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499248 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-run\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499290 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-lib-modules\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499320 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-iscsi\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499327 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-httpd-run\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499387 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-var-locks-brick\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.499422 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.506106 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-scripts\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.509213 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-config-data\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.520297 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.520690 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.521304 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scxgw\" (UniqueName: \"kubernetes.io/projected/6e2d74c2-ac47-4731-a3e5-641274858dad-kube-api-access-scxgw\") pod \"glance-default-internal-api-1\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:47 crc kubenswrapper[4956]: I1126 17:14:47.568986 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:48 crc kubenswrapper[4956]: I1126 17:14:48.100770 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:14:48 crc kubenswrapper[4956]: I1126 17:14:48.171450 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"6e2d74c2-ac47-4731-a3e5-641274858dad","Type":"ContainerStarted","Data":"94beab389dc1be58f65cd7776bc273d9d8b0b433f2a43c4b903c0a1d03d73f63"} Nov 26 17:14:49 crc kubenswrapper[4956]: I1126 17:14:49.008221 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5a647b90-78f6-42fe-a636-493b9ead83b9" path="/var/lib/kubelet/pods/5a647b90-78f6-42fe-a636-493b9ead83b9/volumes" Nov 26 17:14:49 crc kubenswrapper[4956]: I1126 17:14:49.185134 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"6e2d74c2-ac47-4731-a3e5-641274858dad","Type":"ContainerStarted","Data":"48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c"} Nov 26 17:14:49 crc kubenswrapper[4956]: I1126 17:14:49.185210 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"6e2d74c2-ac47-4731-a3e5-641274858dad","Type":"ContainerStarted","Data":"55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0"} Nov 26 17:14:49 crc kubenswrapper[4956]: I1126 17:14:49.231780 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-1" podStartSLOduration=2.231732342 podStartE2EDuration="2.231732342s" podCreationTimestamp="2025-11-26 17:14:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:14:49.210803453 +0000 UTC m=+1334.906764045" watchObservedRunningTime="2025-11-26 17:14:49.231732342 +0000 UTC m=+1334.927692894" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.490772 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.491622 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.521464 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.538954 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.815915 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.816237 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.844753 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:54 crc kubenswrapper[4956]: I1126 17:14:54.854571 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.237194 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.237499 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.237516 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.237526 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.251178 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.251257 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.284474 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:55 crc kubenswrapper[4956]: I1126 17:14:55.295696 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:56 crc kubenswrapper[4956]: I1126 17:14:56.243539 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:56 crc kubenswrapper[4956]: I1126 17:14:56.244025 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.251718 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.251757 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.325055 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.326085 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.384291 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.384476 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.389822 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.498061 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.569703 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.569772 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.602944 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:57 crc kubenswrapper[4956]: I1126 17:14:57.616987 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.260856 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.261406 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.261813 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-log" containerID="cri-o://78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58" gracePeriod=30 Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.262124 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-httpd" containerID="cri-o://40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b" gracePeriod=30 Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.262958 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.262994 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.268710 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.130:9292/healthcheck\": EOF" Nov 26 17:14:58 crc kubenswrapper[4956]: I1126 17:14:58.271458 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.130:9292/healthcheck\": EOF" Nov 26 17:14:59 crc kubenswrapper[4956]: I1126 17:14:59.271757 4956 generic.go:334] "Generic (PLEG): container finished" podID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerID="78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58" exitCode=143 Nov 26 17:14:59 crc kubenswrapper[4956]: I1126 17:14:59.271847 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c","Type":"ContainerDied","Data":"78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58"} Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.157097 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9"] Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.158244 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.163566 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.164832 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.171807 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9"] Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.249953 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-secret-volume\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.250044 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjx75\" (UniqueName: \"kubernetes.io/projected/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-kube-api-access-fjx75\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.250070 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-config-volume\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.351392 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-config-volume\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.351539 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-secret-volume\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.351599 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjx75\" (UniqueName: \"kubernetes.io/projected/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-kube-api-access-fjx75\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.353216 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-config-volume\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.375422 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-secret-volume\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.384224 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjx75\" (UniqueName: \"kubernetes.io/projected/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-kube-api-access-fjx75\") pod \"collect-profiles-29402955-vkvb9\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.481775 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.535347 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.535487 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.552698 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.615410 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.616579 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-httpd" containerID="cri-o://a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff" gracePeriod=30 Nov 26 17:15:00 crc kubenswrapper[4956]: I1126 17:15:00.616977 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-log" containerID="cri-o://da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be" gracePeriod=30 Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.009270 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9"] Nov 26 17:15:01 crc kubenswrapper[4956]: W1126 17:15:01.014815 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod8bcbb3f2_17b2_4856_8bff_b5356eab2cca.slice/crio-b650008ffe237f7079fdc3c0c8ea8e23ee9e61071ecfafe75d957490de7cce6b WatchSource:0}: Error finding container b650008ffe237f7079fdc3c0c8ea8e23ee9e61071ecfafe75d957490de7cce6b: Status 404 returned error can't find the container with id b650008ffe237f7079fdc3c0c8ea8e23ee9e61071ecfafe75d957490de7cce6b Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.296993 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" event={"ID":"8bcbb3f2-17b2-4856-8bff-b5356eab2cca","Type":"ContainerStarted","Data":"4668237196bda52327527f48ecb2eeeab95d312e365380cca6e37b9fca22df9e"} Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.297628 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" event={"ID":"8bcbb3f2-17b2-4856-8bff-b5356eab2cca","Type":"ContainerStarted","Data":"b650008ffe237f7079fdc3c0c8ea8e23ee9e61071ecfafe75d957490de7cce6b"} Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.299634 4956 generic.go:334] "Generic (PLEG): container finished" podID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerID="da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be" exitCode=143 Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.313531 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8","Type":"ContainerDied","Data":"da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be"} Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.340738 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" podStartSLOduration=1.3407096680000001 podStartE2EDuration="1.340709668s" podCreationTimestamp="2025-11-26 17:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:15:01.33457417 +0000 UTC m=+1347.030534722" watchObservedRunningTime="2025-11-26 17:15:01.340709668 +0000 UTC m=+1347.036670220" Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.671378 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-log" probeResult="failure" output="Get \"http://10.217.0.130:9292/healthcheck\": read tcp 10.217.0.2:59306->10.217.0.130:9292: read: connection reset by peer" Nov 26 17:15:01 crc kubenswrapper[4956]: I1126 17:15:01.673425 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.130:9292/healthcheck\": read tcp 10.217.0.2:59314->10.217.0.130:9292: read: connection reset by peer" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.111717 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.187564 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-run\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.188092 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-var-locks-brick\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.189258 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-run" (OuterVolumeSpecName: "run") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.189402 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.189751 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-httpd-run\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.190395 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.190574 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-sys\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.190681 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-config-data\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.190828 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-lib-modules\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.190937 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-nvme\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191098 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-iscsi\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191233 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-dev\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191329 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191433 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-logs\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191552 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2m5w\" (UniqueName: \"kubernetes.io/projected/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-kube-api-access-l2m5w\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191641 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-scripts\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.191729 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\" (UID: \"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c\") " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.192337 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.192432 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.192500 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.193549 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.193587 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-sys" (OuterVolumeSpecName: "sys") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.194371 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-logs" (OuterVolumeSpecName: "logs") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.194445 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.194541 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-dev" (OuterVolumeSpecName: "dev") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.194609 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.200768 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-kube-api-access-l2m5w" (OuterVolumeSpecName: "kube-api-access-l2m5w") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "kube-api-access-l2m5w". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.209848 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.210689 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-scripts" (OuterVolumeSpecName: "scripts") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.219459 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.246230 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-config-data" (OuterVolumeSpecName: "config-data") pod "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" (UID: "4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294018 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294065 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294118 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294131 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294145 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2m5w\" (UniqueName: \"kubernetes.io/projected/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-kube-api-access-l2m5w\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294157 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294181 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294194 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294241 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294258 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.294271 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.308639 4956 generic.go:334] "Generic (PLEG): container finished" podID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerID="40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b" exitCode=0 Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.308714 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c","Type":"ContainerDied","Data":"40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b"} Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.308748 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c","Type":"ContainerDied","Data":"dc016d71d7127772cd969e77fb5258bbf832ad9cf9fd2991c949d5e898327ced"} Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.308766 4956 scope.go:117] "RemoveContainer" containerID="40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.308917 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.311467 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.312130 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.320219 4956 generic.go:334] "Generic (PLEG): container finished" podID="8bcbb3f2-17b2-4856-8bff-b5356eab2cca" containerID="4668237196bda52327527f48ecb2eeeab95d312e365380cca6e37b9fca22df9e" exitCode=0 Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.320286 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" event={"ID":"8bcbb3f2-17b2-4856-8bff-b5356eab2cca","Type":"ContainerDied","Data":"4668237196bda52327527f48ecb2eeeab95d312e365380cca6e37b9fca22df9e"} Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.346408 4956 scope.go:117] "RemoveContainer" containerID="78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.379736 4956 scope.go:117] "RemoveContainer" containerID="40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b" Nov 26 17:15:02 crc kubenswrapper[4956]: E1126 17:15:02.380675 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b\": container with ID starting with 40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b not found: ID does not exist" containerID="40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.384347 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.381380 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b"} err="failed to get container status \"40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b\": rpc error: code = NotFound desc = could not find container \"40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b\": container with ID starting with 40b32cc54352e4e860921cfc72e529efdc685fdb13fff4a7d13c77c33f23452b not found: ID does not exist" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.385894 4956 scope.go:117] "RemoveContainer" containerID="78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58" Nov 26 17:15:02 crc kubenswrapper[4956]: E1126 17:15:02.386686 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58\": container with ID starting with 78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58 not found: ID does not exist" containerID="78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.386730 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58"} err="failed to get container status \"78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58\": rpc error: code = NotFound desc = could not find container \"78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58\": container with ID starting with 78c4ee67773bce6ef597d49fe06fd853d25b1b99e637ebe16dad1caf7eef2d58 not found: ID does not exist" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.396332 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.396388 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.406498 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.415008 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:15:02 crc kubenswrapper[4956]: E1126 17:15:02.415613 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-log" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.415647 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-log" Nov 26 17:15:02 crc kubenswrapper[4956]: E1126 17:15:02.415707 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-httpd" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.415718 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-httpd" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.415956 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-log" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.415981 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" containerName="glance-httpd" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.422448 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.425522 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.497924 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-run\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498010 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498057 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498092 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498126 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498148 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-config-data\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498171 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498256 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498278 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b8g6k\" (UniqueName: \"kubernetes.io/projected/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-kube-api-access-b8g6k\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498329 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498351 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-logs\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498412 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-scripts\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498438 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-sys\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.498455 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-dev\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601261 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601346 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601417 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601464 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601500 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-config-data\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601547 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601578 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601626 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b8g6k\" (UniqueName: \"kubernetes.io/projected/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-kube-api-access-b8g6k\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601667 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601722 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-logs\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601751 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-scripts\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601806 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-sys\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601836 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-dev\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.601903 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-run\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.602039 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-run\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.602384 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") device mount path \"/mnt/openstack/pv09\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605108 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-sys\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605201 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-dev\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605257 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-var-locks-brick\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605318 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-iscsi\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605364 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-nvme\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605336 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605322 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-lib-modules\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605823 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.605825 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-logs\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.611018 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-config-data\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.611365 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-scripts\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.630341 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.635689 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.638488 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b8g6k\" (UniqueName: \"kubernetes.io/projected/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-kube-api-access-b8g6k\") pod \"glance-default-external-api-0\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:02 crc kubenswrapper[4956]: I1126 17:15:02.744701 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.007691 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c" path="/var/lib/kubelet/pods/4c38ee74-c6a3-4ea1-9285-bc85ec9f1b3c/volumes" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.288228 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:15:03 crc kubenswrapper[4956]: W1126 17:15:03.293232 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecee33ed_a8b9_4c53_8e25_311fe5f9757d.slice/crio-7e23315fed85856aca5682c40b49b801df76de548dc471c83d2b45aa272ca114 WatchSource:0}: Error finding container 7e23315fed85856aca5682c40b49b801df76de548dc471c83d2b45aa272ca114: Status 404 returned error can't find the container with id 7e23315fed85856aca5682c40b49b801df76de548dc471c83d2b45aa272ca114 Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.338898 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"ecee33ed-a8b9-4c53-8e25-311fe5f9757d","Type":"ContainerStarted","Data":"7e23315fed85856aca5682c40b49b801df76de548dc471c83d2b45aa272ca114"} Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.626056 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.737413 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjx75\" (UniqueName: \"kubernetes.io/projected/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-kube-api-access-fjx75\") pod \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.737806 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-config-volume\") pod \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.737921 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-secret-volume\") pod \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\" (UID: \"8bcbb3f2-17b2-4856-8bff-b5356eab2cca\") " Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.739219 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-config-volume" (OuterVolumeSpecName: "config-volume") pod "8bcbb3f2-17b2-4856-8bff-b5356eab2cca" (UID: "8bcbb3f2-17b2-4856-8bff-b5356eab2cca"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.741990 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-kube-api-access-fjx75" (OuterVolumeSpecName: "kube-api-access-fjx75") pod "8bcbb3f2-17b2-4856-8bff-b5356eab2cca" (UID: "8bcbb3f2-17b2-4856-8bff-b5356eab2cca"). InnerVolumeSpecName "kube-api-access-fjx75". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.744084 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "8bcbb3f2-17b2-4856-8bff-b5356eab2cca" (UID: "8bcbb3f2-17b2-4856-8bff-b5356eab2cca"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.840497 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjx75\" (UniqueName: \"kubernetes.io/projected/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-kube-api-access-fjx75\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.841073 4956 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-config-volume\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:03 crc kubenswrapper[4956]: I1126 17:15:03.841090 4956 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/8bcbb3f2-17b2-4856-8bff-b5356eab2cca-secret-volume\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.111024 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148254 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-lib-modules\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148339 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-iscsi\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148405 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-scripts\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148436 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-nvme\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148445 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148523 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148548 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-sys\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148583 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-dev\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148595 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148606 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-sys" (OuterVolumeSpecName: "sys") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148645 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148693 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-run\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148735 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2pvp7\" (UniqueName: \"kubernetes.io/projected/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-kube-api-access-2pvp7\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148760 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-dev" (OuterVolumeSpecName: "dev") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148825 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148792 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-run" (OuterVolumeSpecName: "run") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.148786 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-var-locks-brick\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.149042 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.149144 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-config-data\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.149173 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-logs\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.149236 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-httpd-run\") pod \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\" (UID: \"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8\") " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.150220 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-logs" (OuterVolumeSpecName: "logs") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.150581 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151018 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151043 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151058 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151071 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151085 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151097 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151111 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151123 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.151135 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.156043 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-kube-api-access-2pvp7" (OuterVolumeSpecName: "kube-api-access-2pvp7") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "kube-api-access-2pvp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.158636 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.158796 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-scripts" (OuterVolumeSpecName: "scripts") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.160524 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance-cache") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.201124 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-config-data" (OuterVolumeSpecName: "config-data") pod "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" (UID: "b6c1ed78-2995-4a5f-9bf4-d489f37f55d8"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.253357 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.253824 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2pvp7\" (UniqueName: \"kubernetes.io/projected/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-kube-api-access-2pvp7\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.253849 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.253858 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.253888 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.269362 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.271523 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.355537 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.355595 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.366458 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" event={"ID":"8bcbb3f2-17b2-4856-8bff-b5356eab2cca","Type":"ContainerDied","Data":"b650008ffe237f7079fdc3c0c8ea8e23ee9e61071ecfafe75d957490de7cce6b"} Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.366514 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b650008ffe237f7079fdc3c0c8ea8e23ee9e61071ecfafe75d957490de7cce6b" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.366609 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29402955-vkvb9" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.391214 4956 generic.go:334] "Generic (PLEG): container finished" podID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerID="a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff" exitCode=0 Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.391291 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8","Type":"ContainerDied","Data":"a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff"} Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.391329 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"b6c1ed78-2995-4a5f-9bf4-d489f37f55d8","Type":"ContainerDied","Data":"201e3b2cfff8b4f2cb63208e56e0e55a0371354103bec5152762fc678a97bf79"} Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.391347 4956 scope.go:117] "RemoveContainer" containerID="a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.391490 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.400895 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"ecee33ed-a8b9-4c53-8e25-311fe5f9757d","Type":"ContainerStarted","Data":"539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae"} Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.400944 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"ecee33ed-a8b9-4c53-8e25-311fe5f9757d","Type":"ContainerStarted","Data":"656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4"} Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.434634 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-external-api-0" podStartSLOduration=2.4346119059999998 podStartE2EDuration="2.434611906s" podCreationTimestamp="2025-11-26 17:15:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:15:04.432180085 +0000 UTC m=+1350.128140637" watchObservedRunningTime="2025-11-26 17:15:04.434611906 +0000 UTC m=+1350.130572458" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.462490 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.469678 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.474897 4956 scope.go:117] "RemoveContainer" containerID="da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.488400 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:15:04 crc kubenswrapper[4956]: E1126 17:15:04.488777 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8bcbb3f2-17b2-4856-8bff-b5356eab2cca" containerName="collect-profiles" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.488798 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="8bcbb3f2-17b2-4856-8bff-b5356eab2cca" containerName="collect-profiles" Nov 26 17:15:04 crc kubenswrapper[4956]: E1126 17:15:04.488822 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-log" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.488831 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-log" Nov 26 17:15:04 crc kubenswrapper[4956]: E1126 17:15:04.488844 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-httpd" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.488851 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-httpd" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.489002 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-log" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.489017 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" containerName="glance-httpd" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.489035 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="8bcbb3f2-17b2-4856-8bff-b5356eab2cca" containerName="collect-profiles" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.490127 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.504822 4956 scope.go:117] "RemoveContainer" containerID="a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff" Nov 26 17:15:04 crc kubenswrapper[4956]: E1126 17:15:04.510200 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff\": container with ID starting with a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff not found: ID does not exist" containerID="a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.510313 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff"} err="failed to get container status \"a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff\": rpc error: code = NotFound desc = could not find container \"a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff\": container with ID starting with a91a9177e3497be5b05c03d8f75cb5ef389a7a770b29ff8c14beae62f394cfff not found: ID does not exist" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.510360 4956 scope.go:117] "RemoveContainer" containerID="da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be" Nov 26 17:15:04 crc kubenswrapper[4956]: E1126 17:15:04.512492 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be\": container with ID starting with da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be not found: ID does not exist" containerID="da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.512529 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be"} err="failed to get container status \"da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be\": rpc error: code = NotFound desc = could not find container \"da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be\": container with ID starting with da2953b4a6d34a730b0e91f723fe2340cde75457794ad72f0e7163554dcf92be not found: ID does not exist" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.514447 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.558849 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559068 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559107 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559132 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-logs\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559307 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559343 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-sys\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559376 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559398 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559418 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559457 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-run\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559551 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559586 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wps7x\" (UniqueName: \"kubernetes.io/projected/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-kube-api-access-wps7x\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559721 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-dev\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.559779 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.661558 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.661640 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.661799 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.661756 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-nvme\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.661824 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-logs\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.661979 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662019 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-sys\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662105 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662165 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") device mount path \"/mnt/openstack/pv03\"" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662230 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-sys\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662340 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662386 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662385 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-iscsi\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662412 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662549 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-run\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662605 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662631 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wps7x\" (UniqueName: \"kubernetes.io/projected/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-kube-api-access-wps7x\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662625 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-lib-modules\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662666 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-dev\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662713 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-dev\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662726 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-run\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662894 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.662979 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-var-locks-brick\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.663200 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-logs\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.663242 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.677845 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-config-data\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.686687 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wps7x\" (UniqueName: \"kubernetes.io/projected/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-kube-api-access-wps7x\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.687747 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-scripts\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.689668 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.712125 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"glance-default-internal-api-0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:04 crc kubenswrapper[4956]: I1126 17:15:04.813742 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:05 crc kubenswrapper[4956]: I1126 17:15:05.008076 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6c1ed78-2995-4a5f-9bf4-d489f37f55d8" path="/var/lib/kubelet/pods/b6c1ed78-2995-4a5f-9bf4-d489f37f55d8/volumes" Nov 26 17:15:05 crc kubenswrapper[4956]: I1126 17:15:05.097339 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:15:05 crc kubenswrapper[4956]: I1126 17:15:05.416280 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0","Type":"ContainerStarted","Data":"14bb89712c231581faf3aed7609ee42ef2db23b5b193c548c3bb5c51efc45860"} Nov 26 17:15:07 crc kubenswrapper[4956]: I1126 17:15:07.441012 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0","Type":"ContainerStarted","Data":"5fcb8669653662238df17c20f64712eae066c96abd77f821c064a0be03eed32a"} Nov 26 17:15:07 crc kubenswrapper[4956]: I1126 17:15:07.442106 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0","Type":"ContainerStarted","Data":"8b7ae5f26d68354fc34f1d985327638a2c9771cb3e36ad8276954a7d5f77a48d"} Nov 26 17:15:09 crc kubenswrapper[4956]: I1126 17:15:09.551439 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:15:09 crc kubenswrapper[4956]: I1126 17:15:09.551999 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:15:12 crc kubenswrapper[4956]: I1126 17:15:12.744954 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:12 crc kubenswrapper[4956]: I1126 17:15:12.745444 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:12 crc kubenswrapper[4956]: I1126 17:15:12.773674 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:12 crc kubenswrapper[4956]: I1126 17:15:12.797564 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-internal-api-0" podStartSLOduration=8.797542862 podStartE2EDuration="8.797542862s" podCreationTimestamp="2025-11-26 17:15:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:15:07.478460176 +0000 UTC m=+1353.174420748" watchObservedRunningTime="2025-11-26 17:15:12.797542862 +0000 UTC m=+1358.493503414" Nov 26 17:15:12 crc kubenswrapper[4956]: I1126 17:15:12.799784 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:13 crc kubenswrapper[4956]: I1126 17:15:13.495431 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:13 crc kubenswrapper[4956]: I1126 17:15:13.495490 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:14 crc kubenswrapper[4956]: I1126 17:15:14.814369 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:14 crc kubenswrapper[4956]: I1126 17:15:14.816128 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:14 crc kubenswrapper[4956]: I1126 17:15:14.840176 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:14 crc kubenswrapper[4956]: I1126 17:15:14.852700 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:15 crc kubenswrapper[4956]: I1126 17:15:15.513707 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:15 crc kubenswrapper[4956]: I1126 17:15:15.513951 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:15 crc kubenswrapper[4956]: I1126 17:15:15.645784 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:15 crc kubenswrapper[4956]: I1126 17:15:15.646363 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:15:15 crc kubenswrapper[4956]: I1126 17:15:15.655584 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:15:17 crc kubenswrapper[4956]: I1126 17:15:17.574288 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:17 crc kubenswrapper[4956]: I1126 17:15:17.575408 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:15:17 crc kubenswrapper[4956]: I1126 17:15:17.637399 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.551612 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.553125 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.553219 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.554118 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6afa66b60ccff62ba0cee9768e910db0d5a65279fb098588258f5b5dc187f77e"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.554178 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://6afa66b60ccff62ba0cee9768e910db0d5a65279fb098588258f5b5dc187f77e" gracePeriod=600 Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.737642 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="6afa66b60ccff62ba0cee9768e910db0d5a65279fb098588258f5b5dc187f77e" exitCode=0 Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.737697 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"6afa66b60ccff62ba0cee9768e910db0d5a65279fb098588258f5b5dc187f77e"} Nov 26 17:15:39 crc kubenswrapper[4956]: I1126 17:15:39.737743 4956 scope.go:117] "RemoveContainer" containerID="e26914562ac9efcf004522ee5db12eb4d489e28d6cbf474316c25c32bd557c51" Nov 26 17:15:40 crc kubenswrapper[4956]: I1126 17:15:40.749526 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186"} Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.006388 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.008241 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.017327 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.017659 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.022719 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.098945 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.099009 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.200923 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.200999 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.201130 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.220940 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.364655 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:47 crc kubenswrapper[4956]: I1126 17:15:47.849739 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Nov 26 17:15:48 crc kubenswrapper[4956]: I1126 17:15:48.817628 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e47afe75-ccd9-48fa-81e8-ea9bf59826fc","Type":"ContainerStarted","Data":"86702efe1fc04bab2e1672488b41f66dbd1ce41438821c22c919bfbcf5e7e611"} Nov 26 17:15:48 crc kubenswrapper[4956]: I1126 17:15:48.818190 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e47afe75-ccd9-48fa-81e8-ea9bf59826fc","Type":"ContainerStarted","Data":"87c188c3104c680e689db9de412d5e56a39272d6598f29fa49e1015007160e08"} Nov 26 17:15:48 crc kubenswrapper[4956]: I1126 17:15:48.841120 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-9-crc" podStartSLOduration=2.841099118 podStartE2EDuration="2.841099118s" podCreationTimestamp="2025-11-26 17:15:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:15:48.833834536 +0000 UTC m=+1394.529795108" watchObservedRunningTime="2025-11-26 17:15:48.841099118 +0000 UTC m=+1394.537059670" Nov 26 17:15:49 crc kubenswrapper[4956]: I1126 17:15:49.830012 4956 generic.go:334] "Generic (PLEG): container finished" podID="e47afe75-ccd9-48fa-81e8-ea9bf59826fc" containerID="86702efe1fc04bab2e1672488b41f66dbd1ce41438821c22c919bfbcf5e7e611" exitCode=0 Nov 26 17:15:49 crc kubenswrapper[4956]: I1126 17:15:49.830112 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e47afe75-ccd9-48fa-81e8-ea9bf59826fc","Type":"ContainerDied","Data":"86702efe1fc04bab2e1672488b41f66dbd1ce41438821c22c919bfbcf5e7e611"} Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.169262 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.293225 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kube-api-access\") pod \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.293291 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kubelet-dir\") pod \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\" (UID: \"e47afe75-ccd9-48fa-81e8-ea9bf59826fc\") " Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.293779 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "e47afe75-ccd9-48fa-81e8-ea9bf59826fc" (UID: "e47afe75-ccd9-48fa-81e8-ea9bf59826fc"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.300269 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e47afe75-ccd9-48fa-81e8-ea9bf59826fc" (UID: "e47afe75-ccd9-48fa-81e8-ea9bf59826fc"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.395616 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.395681 4956 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/e47afe75-ccd9-48fa-81e8-ea9bf59826fc-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.807130 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 17:15:51 crc kubenswrapper[4956]: E1126 17:15:51.807486 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e47afe75-ccd9-48fa-81e8-ea9bf59826fc" containerName="pruner" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.807508 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="e47afe75-ccd9-48fa-81e8-ea9bf59826fc" containerName="pruner" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.807656 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="e47afe75-ccd9-48fa-81e8-ea9bf59826fc" containerName="pruner" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.808843 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.820468 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.858494 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"e47afe75-ccd9-48fa-81e8-ea9bf59826fc","Type":"ContainerDied","Data":"87c188c3104c680e689db9de412d5e56a39272d6598f29fa49e1015007160e08"} Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.858566 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="87c188c3104c680e689db9de412d5e56a39272d6598f29fa49e1015007160e08" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.858662 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.904405 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-var-lock\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.904483 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:51 crc kubenswrapper[4956]: I1126 17:15:51.904546 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.005982 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-var-lock\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.006047 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.006097 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.006474 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-var-lock\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.006492 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kubelet-dir\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.025854 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kube-api-access\") pod \"installer-9-crc\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.147622 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.578463 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Nov 26 17:15:52 crc kubenswrapper[4956]: I1126 17:15:52.868452 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d72606f-03e5-4a3a-a98e-05e56db0d2eb","Type":"ContainerStarted","Data":"f990776e5ff12c2695a1a4ca7a44c87725040c74f072ffaab3b3c711d94c6ea1"} Nov 26 17:15:53 crc kubenswrapper[4956]: I1126 17:15:53.888998 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d72606f-03e5-4a3a-a98e-05e56db0d2eb","Type":"ContainerStarted","Data":"d0abb6914ca35eea14580a7958053585360705a7c300d2afa7c1cc86251dfed2"} Nov 26 17:15:53 crc kubenswrapper[4956]: I1126 17:15:53.921940 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=2.921911893 podStartE2EDuration="2.921911893s" podCreationTimestamp="2025-11-26 17:15:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:15:53.914762215 +0000 UTC m=+1399.610722777" watchObservedRunningTime="2025-11-26 17:15:53.921911893 +0000 UTC m=+1399.617872465" Nov 26 17:16:07 crc kubenswrapper[4956]: I1126 17:16:07.952317 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:16:07 crc kubenswrapper[4956]: I1126 17:16:07.955022 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-log" containerID="cri-o://5767b20b397a28d1c45dcf61e4de7c5a4a7f897fce042c2a43444107bb4e6e59" gracePeriod=30 Nov 26 17:16:07 crc kubenswrapper[4956]: I1126 17:16:07.955206 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-1" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-httpd" containerID="cri-o://18551c8bd2a6a38f12abf66c667eb04138078566eef08a87c1dc50ecb7afa55d" gracePeriod=30 Nov 26 17:16:08 crc kubenswrapper[4956]: I1126 17:16:08.179913 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:16:08 crc kubenswrapper[4956]: I1126 17:16:08.180251 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-log" containerID="cri-o://55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0" gracePeriod=30 Nov 26 17:16:08 crc kubenswrapper[4956]: I1126 17:16:08.180406 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-1" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-httpd" containerID="cri-o://48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c" gracePeriod=30 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.031696 4956 generic.go:334] "Generic (PLEG): container finished" podID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerID="55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0" exitCode=143 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.031814 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"6e2d74c2-ac47-4731-a3e5-641274858dad","Type":"ContainerDied","Data":"55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0"} Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.034804 4956 generic.go:334] "Generic (PLEG): container finished" podID="e164968b-dfde-4d68-b73e-632722876b76" containerID="5767b20b397a28d1c45dcf61e4de7c5a4a7f897fce042c2a43444107bb4e6e59" exitCode=143 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.034832 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e164968b-dfde-4d68-b73e-632722876b76","Type":"ContainerDied","Data":"5767b20b397a28d1c45dcf61e4de7c5a4a7f897fce042c2a43444107bb4e6e59"} Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.381630 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x9hlb"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.387927 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-x9hlb"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.469460 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.469802 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-log" containerID="cri-o://656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4" gracePeriod=30 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.469919 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-external-api-0" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-httpd" containerID="cri-o://539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae" gracePeriod=30 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.477888 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glanced650-account-delete-bjkfl"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.479118 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.491378 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced650-account-delete-bjkfl"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.537068 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0267ab5c-7b30-465a-89e3-67af886a3348-operator-scripts\") pod \"glanced650-account-delete-bjkfl\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.537149 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5rggd\" (UniqueName: \"kubernetes.io/projected/0267ab5c-7b30-465a-89e3-67af886a3348-kube-api-access-5rggd\") pod \"glanced650-account-delete-bjkfl\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.561713 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.562065 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-log" containerID="cri-o://8b7ae5f26d68354fc34f1d985327638a2c9771cb3e36ad8276954a7d5f77a48d" gracePeriod=30 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.562178 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-internal-api-0" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-httpd" containerID="cri-o://5fcb8669653662238df17c20f64712eae066c96abd77f821c064a0be03eed32a" gracePeriod=30 Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.638964 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0267ab5c-7b30-465a-89e3-67af886a3348-operator-scripts\") pod \"glanced650-account-delete-bjkfl\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.639452 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5rggd\" (UniqueName: \"kubernetes.io/projected/0267ab5c-7b30-465a-89e3-67af886a3348-kube-api-access-5rggd\") pod \"glanced650-account-delete-bjkfl\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.640029 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0267ab5c-7b30-465a-89e3-67af886a3348-operator-scripts\") pod \"glanced650-account-delete-bjkfl\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.676859 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5rggd\" (UniqueName: \"kubernetes.io/projected/0267ab5c-7b30-465a-89e3-67af886a3348-kube-api-access-5rggd\") pod \"glanced650-account-delete-bjkfl\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.800649 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.845020 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:16:09 crc kubenswrapper[4956]: I1126 17:16:09.845448 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/openstackclient" podUID="d473f051-a35a-4c66-86ac-21c54348895e" containerName="openstackclient" containerID="cri-o://2f2e7e1f79b9f23f1c20072ce2f9071c04f6bf83c98f34ffab3f58015fe36aa5" gracePeriod=30 Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.066836 4956 generic.go:334] "Generic (PLEG): container finished" podID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerID="8b7ae5f26d68354fc34f1d985327638a2c9771cb3e36ad8276954a7d5f77a48d" exitCode=143 Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.066972 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0","Type":"ContainerDied","Data":"8b7ae5f26d68354fc34f1d985327638a2c9771cb3e36ad8276954a7d5f77a48d"} Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.076294 4956 generic.go:334] "Generic (PLEG): container finished" podID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerID="656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4" exitCode=143 Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.076379 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"ecee33ed-a8b9-4c53-8e25-311fe5f9757d","Type":"ContainerDied","Data":"656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4"} Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.078527 4956 generic.go:334] "Generic (PLEG): container finished" podID="d473f051-a35a-4c66-86ac-21c54348895e" containerID="2f2e7e1f79b9f23f1c20072ce2f9071c04f6bf83c98f34ffab3f58015fe36aa5" exitCode=143 Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.078562 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d473f051-a35a-4c66-86ac-21c54348895e","Type":"ContainerDied","Data":"2f2e7e1f79b9f23f1c20072ce2f9071c04f6bf83c98f34ffab3f58015fe36aa5"} Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.299665 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glanced650-account-delete-bjkfl"] Nov 26 17:16:10 crc kubenswrapper[4956]: W1126 17:16:10.310986 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0267ab5c_7b30_465a_89e3_67af886a3348.slice/crio-188614107c454e3548120cd59345f0da9b835fba46179eaa93e20893450086d2 WatchSource:0}: Error finding container 188614107c454e3548120cd59345f0da9b835fba46179eaa93e20893450086d2: Status 404 returned error can't find the container with id 188614107c454e3548120cd59345f0da9b835fba46179eaa93e20893450086d2 Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.427373 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.559168 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-scripts\") pod \"d473f051-a35a-4c66-86ac-21c54348895e\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.559365 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config\") pod \"d473f051-a35a-4c66-86ac-21c54348895e\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.559416 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config-secret\") pod \"d473f051-a35a-4c66-86ac-21c54348895e\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.559656 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6z2jp\" (UniqueName: \"kubernetes.io/projected/d473f051-a35a-4c66-86ac-21c54348895e-kube-api-access-6z2jp\") pod \"d473f051-a35a-4c66-86ac-21c54348895e\" (UID: \"d473f051-a35a-4c66-86ac-21c54348895e\") " Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.560213 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-scripts" (OuterVolumeSpecName: "openstack-scripts") pod "d473f051-a35a-4c66-86ac-21c54348895e" (UID: "d473f051-a35a-4c66-86ac-21c54348895e"). InnerVolumeSpecName "openstack-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.568929 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d473f051-a35a-4c66-86ac-21c54348895e-kube-api-access-6z2jp" (OuterVolumeSpecName: "kube-api-access-6z2jp") pod "d473f051-a35a-4c66-86ac-21c54348895e" (UID: "d473f051-a35a-4c66-86ac-21c54348895e"). InnerVolumeSpecName "kube-api-access-6z2jp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.583916 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config" (OuterVolumeSpecName: "openstack-config") pod "d473f051-a35a-4c66-86ac-21c54348895e" (UID: "d473f051-a35a-4c66-86ac-21c54348895e"). InnerVolumeSpecName "openstack-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.586104 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config-secret" (OuterVolumeSpecName: "openstack-config-secret") pod "d473f051-a35a-4c66-86ac-21c54348895e" (UID: "d473f051-a35a-4c66-86ac-21c54348895e"). InnerVolumeSpecName "openstack-config-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.666543 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6z2jp\" (UniqueName: \"kubernetes.io/projected/d473f051-a35a-4c66-86ac-21c54348895e-kube-api-access-6z2jp\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.666597 4956 reconciler_common.go:293] "Volume detached for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.666617 4956 reconciler_common.go:293] "Volume detached for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:10 crc kubenswrapper[4956]: I1126 17:16:10.666631 4956 reconciler_common.go:293] "Volume detached for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/d473f051-a35a-4c66-86ac-21c54348895e-openstack-config-secret\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.005945 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0037dfab-b9ac-4c2f-9f17-170f34a60b5a" path="/var/lib/kubelet/pods/0037dfab-b9ac-4c2f-9f17-170f34a60b5a/volumes" Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.121818 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"d473f051-a35a-4c66-86ac-21c54348895e","Type":"ContainerDied","Data":"c50b6e57c05a9d8c1e86688910820e795f35843add1cc754dd29631a5aa63a92"} Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.122302 4956 scope.go:117] "RemoveContainer" containerID="2f2e7e1f79b9f23f1c20072ce2f9071c04f6bf83c98f34ffab3f58015fe36aa5" Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.122485 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.126704 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" event={"ID":"0267ab5c-7b30-465a-89e3-67af886a3348","Type":"ContainerStarted","Data":"46b2afe6fccde0256021d6e3ddf58a8ed8a149d4b75752394c111b7e15c674f5"} Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.126743 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" event={"ID":"0267ab5c-7b30-465a-89e3-67af886a3348","Type":"ContainerStarted","Data":"188614107c454e3548120cd59345f0da9b835fba46179eaa93e20893450086d2"} Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.154330 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.162133 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.170434 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" podStartSLOduration=2.170395822 podStartE2EDuration="2.170395822s" podCreationTimestamp="2025-11-26 17:16:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:16:11.170313549 +0000 UTC m=+1416.866274121" watchObservedRunningTime="2025-11-26 17:16:11.170395822 +0000 UTC m=+1416.866356374" Nov 26 17:16:11 crc kubenswrapper[4956]: I1126 17:16:11.865624 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.029789 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-nvme\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.029930 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.030250 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.030390 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-dev" (OuterVolumeSpecName: "dev") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.030939 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-dev\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031087 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-logs\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031540 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-iscsi\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031716 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-sys\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031898 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-lib-modules\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.032066 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.032563 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-run\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.032689 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-httpd-run\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031503 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-logs" (OuterVolumeSpecName: "logs") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031630 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031824 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-sys" (OuterVolumeSpecName: "sys") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.031978 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.032671 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-run" (OuterVolumeSpecName: "run") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.032945 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.033191 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-scripts\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.033307 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-config-data\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.033399 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-var-locks-brick\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.033492 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scxgw\" (UniqueName: \"kubernetes.io/projected/6e2d74c2-ac47-4731-a3e5-641274858dad-kube-api-access-scxgw\") pod \"6e2d74c2-ac47-4731-a3e5-641274858dad\" (UID: \"6e2d74c2-ac47-4731-a3e5-641274858dad\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.033493 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034254 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034323 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034386 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034443 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034494 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034543 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034592 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/6e2d74c2-ac47-4731-a3e5-641274858dad-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034649 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.034708 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/6e2d74c2-ac47-4731-a3e5-641274858dad-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.038038 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.038063 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.053687 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6e2d74c2-ac47-4731-a3e5-641274858dad-kube-api-access-scxgw" (OuterVolumeSpecName: "kube-api-access-scxgw") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "kube-api-access-scxgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.055189 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-scripts" (OuterVolumeSpecName: "scripts") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.088051 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-config-data" (OuterVolumeSpecName: "config-data") pod "6e2d74c2-ac47-4731-a3e5-641274858dad" (UID: "6e2d74c2-ac47-4731-a3e5-641274858dad"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.137313 4956 generic.go:334] "Generic (PLEG): container finished" podID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerID="48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c" exitCode=0 Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.137366 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"6e2d74c2-ac47-4731-a3e5-641274858dad","Type":"ContainerDied","Data":"48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c"} Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.138613 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-1" event={"ID":"6e2d74c2-ac47-4731-a3e5-641274858dad","Type":"ContainerDied","Data":"94beab389dc1be58f65cd7776bc273d9d8b0b433f2a43c4b903c0a1d03d73f63"} Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.137482 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.138653 4956 scope.go:117] "RemoveContainer" containerID="48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.138686 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.138707 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6e2d74c2-ac47-4731-a3e5-641274858dad-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.138723 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scxgw\" (UniqueName: \"kubernetes.io/projected/6e2d74c2-ac47-4731-a3e5-641274858dad-kube-api-access-scxgw\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.138766 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.137410 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-1" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.140915 4956 generic.go:334] "Generic (PLEG): container finished" podID="0267ab5c-7b30-465a-89e3-67af886a3348" containerID="46b2afe6fccde0256021d6e3ddf58a8ed8a149d4b75752394c111b7e15c674f5" exitCode=0 Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.141075 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" event={"ID":"0267ab5c-7b30-465a-89e3-67af886a3348","Type":"ContainerDied","Data":"46b2afe6fccde0256021d6e3ddf58a8ed8a149d4b75752394c111b7e15c674f5"} Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.143451 4956 generic.go:334] "Generic (PLEG): container finished" podID="e164968b-dfde-4d68-b73e-632722876b76" containerID="18551c8bd2a6a38f12abf66c667eb04138078566eef08a87c1dc50ecb7afa55d" exitCode=0 Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.143510 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e164968b-dfde-4d68-b73e-632722876b76","Type":"ContainerDied","Data":"18551c8bd2a6a38f12abf66c667eb04138078566eef08a87c1dc50ecb7afa55d"} Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.143527 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-1" event={"ID":"e164968b-dfde-4d68-b73e-632722876b76","Type":"ContainerDied","Data":"92fbaaf108139cb95a45215dbf12639f90a8e4951b69d89701ed22f900013493"} Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.143748 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="92fbaaf108139cb95a45215dbf12639f90a8e4951b69d89701ed22f900013493" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.161851 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.162380 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.165667 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.180017 4956 scope.go:117] "RemoveContainer" containerID="55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.181732 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.190606 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-1"] Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.204975 4956 scope.go:117] "RemoveContainer" containerID="48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c" Nov 26 17:16:12 crc kubenswrapper[4956]: E1126 17:16:12.205542 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c\": container with ID starting with 48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c not found: ID does not exist" containerID="48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.205578 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c"} err="failed to get container status \"48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c\": rpc error: code = NotFound desc = could not find container \"48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c\": container with ID starting with 48538dbfa5b7de54e93998c68f45848ff405a524dc63b4ba8d40ef9818db6b3c not found: ID does not exist" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.205603 4956 scope.go:117] "RemoveContainer" containerID="55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0" Nov 26 17:16:12 crc kubenswrapper[4956]: E1126 17:16:12.206031 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0\": container with ID starting with 55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0 not found: ID does not exist" containerID="55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.206055 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0"} err="failed to get container status \"55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0\": rpc error: code = NotFound desc = could not find container \"55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0\": container with ID starting with 55ec2caa335e3849531f1938928276eeb31b2576fb28d736a31d8d3f038a3ae0 not found: ID does not exist" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.240251 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.240289 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.341991 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-config-data\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342057 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-nvme\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342093 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-dev\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342183 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-lib-modules\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342229 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-logs\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342240 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342285 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-scripts\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342334 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342336 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342413 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-var-locks-brick\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342412 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-dev" (OuterVolumeSpecName: "dev") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342461 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-iscsi\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342508 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cssg6\" (UniqueName: \"kubernetes.io/projected/e164968b-dfde-4d68-b73e-632722876b76-kube-api-access-cssg6\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342510 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342560 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-httpd-run\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342619 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-run\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342665 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-sys\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342705 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"e164968b-dfde-4d68-b73e-632722876b76\" (UID: \"e164968b-dfde-4d68-b73e-632722876b76\") " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342560 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.342923 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-run" (OuterVolumeSpecName: "run") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343000 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-sys" (OuterVolumeSpecName: "sys") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343048 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343369 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-logs" (OuterVolumeSpecName: "logs") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343777 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343801 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343810 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343820 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343828 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343840 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343849 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/e164968b-dfde-4d68-b73e-632722876b76-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343857 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.343880 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/e164968b-dfde-4d68-b73e-632722876b76-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.345606 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "glance-cache") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.346169 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-scripts" (OuterVolumeSpecName: "scripts") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.346171 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage11-crc" (OuterVolumeSpecName: "glance") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "local-storage11-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.346938 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e164968b-dfde-4d68-b73e-632722876b76-kube-api-access-cssg6" (OuterVolumeSpecName: "kube-api-access-cssg6") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "kube-api-access-cssg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.380058 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-config-data" (OuterVolumeSpecName: "config-data") pod "e164968b-dfde-4d68-b73e-632722876b76" (UID: "e164968b-dfde-4d68-b73e-632722876b76"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.445239 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cssg6\" (UniqueName: \"kubernetes.io/projected/e164968b-dfde-4d68-b73e-632722876b76-kube-api-access-cssg6\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.445330 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.445345 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.445359 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e164968b-dfde-4d68-b73e-632722876b76-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.445378 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.459557 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage11-crc" (UniqueName: "kubernetes.io/local-volume/local-storage11-crc") on node "crc" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.461175 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.547291 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:12 crc kubenswrapper[4956]: I1126 17:16:12.547400 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.003389 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" path="/var/lib/kubelet/pods/6e2d74c2-ac47-4731-a3e5-641274858dad/volumes" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.004799 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d473f051-a35a-4c66-86ac-21c54348895e" path="/var/lib/kubelet/pods/d473f051-a35a-4c66-86ac-21c54348895e/volumes" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.104115 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.160915 4956 generic.go:334] "Generic (PLEG): container finished" podID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerID="5fcb8669653662238df17c20f64712eae066c96abd77f821c064a0be03eed32a" exitCode=0 Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.160995 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0","Type":"ContainerDied","Data":"5fcb8669653662238df17c20f64712eae066c96abd77f821c064a0be03eed32a"} Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.161031 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-internal-api-0" event={"ID":"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0","Type":"ContainerDied","Data":"14bb89712c231581faf3aed7609ee42ef2db23b5b193c548c3bb5c51efc45860"} Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.161048 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="14bb89712c231581faf3aed7609ee42ef2db23b5b193c548c3bb5c51efc45860" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.163304 4956 generic.go:334] "Generic (PLEG): container finished" podID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerID="539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae" exitCode=0 Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.163377 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-0" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.163402 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"ecee33ed-a8b9-4c53-8e25-311fe5f9757d","Type":"ContainerDied","Data":"539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae"} Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.163460 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-external-api-0" event={"ID":"ecee33ed-a8b9-4c53-8e25-311fe5f9757d","Type":"ContainerDied","Data":"7e23315fed85856aca5682c40b49b801df76de548dc471c83d2b45aa272ca114"} Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.163486 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-external-api-1" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.163487 4956 scope.go:117] "RemoveContainer" containerID="539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.220827 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.239490 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.251241 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-1"] Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.253981 4956 scope.go:117] "RemoveContainer" containerID="656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261129 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-var-locks-brick\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261222 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-run\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261281 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-httpd-run\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261360 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-logs\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261481 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-config-data\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261533 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261610 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-scripts\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261634 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261687 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-nvme\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.261739 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-iscsi\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262075 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b8g6k\" (UniqueName: \"kubernetes.io/projected/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-kube-api-access-b8g6k\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262104 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-dev\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262142 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-sys\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262165 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-lib-modules\") pod \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\" (UID: \"ecee33ed-a8b9-4c53-8e25-311fe5f9757d\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262679 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262724 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.262762 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-run" (OuterVolumeSpecName: "run") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.263156 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.263497 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-logs" (OuterVolumeSpecName: "logs") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.264479 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.270792 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.274834 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-scripts" (OuterVolumeSpecName: "scripts") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.279042 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage09-crc" (OuterVolumeSpecName: "glance-cache") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "local-storage09-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.283010 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.283102 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-dev" (OuterVolumeSpecName: "dev") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.283487 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-sys" (OuterVolumeSpecName: "sys") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.309592 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-kube-api-access-b8g6k" (OuterVolumeSpecName: "kube-api-access-b8g6k") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "kube-api-access-b8g6k". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.357155 4956 scope.go:117] "RemoveContainer" containerID="539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae" Nov 26 17:16:13 crc kubenswrapper[4956]: E1126 17:16:13.370089 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae\": container with ID starting with 539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae not found: ID does not exist" containerID="539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370176 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae"} err="failed to get container status \"539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae\": rpc error: code = NotFound desc = could not find container \"539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae\": container with ID starting with 539e467e06b11ed18fee35511eaffb1c8ec90079bf7f5c574726e153c935d6ae not found: ID does not exist" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370216 4956 scope.go:117] "RemoveContainer" containerID="656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370633 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-dev\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370762 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-config-data\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370810 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370851 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-var-locks-brick\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.370900 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-scripts\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: E1126 17:16:13.370961 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4\": container with ID starting with 656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4 not found: ID does not exist" containerID="656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371014 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4"} err="failed to get container status \"656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4\": rpc error: code = NotFound desc = could not find container \"656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4\": container with ID starting with 656c713ebf60905617aa2827d240af676b17eb6e549af659ad85cc23076d99d4 not found: ID does not exist" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371120 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-run\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371184 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-lib-modules\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371205 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-sys\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371245 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371300 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-logs\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371323 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wps7x\" (UniqueName: \"kubernetes.io/projected/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-kube-api-access-wps7x\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371353 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-iscsi\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371379 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-httpd-run\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371425 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-nvme\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371790 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371807 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371822 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371835 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371843 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371875 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b8g6k\" (UniqueName: \"kubernetes.io/projected/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-kube-api-access-b8g6k\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371895 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371908 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371917 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371970 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371985 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.371994 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.372002 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.383009 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.385019 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-dev" (OuterVolumeSpecName: "dev") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.385457 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.385490 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-logs" (OuterVolumeSpecName: "logs") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.385585 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-run" (OuterVolumeSpecName: "run") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.385611 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.385636 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-sys" (OuterVolumeSpecName: "sys") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.387219 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.387276 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.388006 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance-cache") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.419628 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-scripts" (OuterVolumeSpecName: "scripts") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.433239 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-kube-api-access-wps7x" (OuterVolumeSpecName: "kube-api-access-wps7x") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "kube-api-access-wps7x". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.452464 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage09-crc" (UniqueName: "kubernetes.io/local-volume/local-storage09-crc") on node "crc" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.485477 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.488540 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") pod \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\" (UID: \"a9ac260d-3fb7-4391-b5c3-bb05d27b79e0\") " Nov 26 17:16:13 crc kubenswrapper[4956]: W1126 17:16:13.488847 4956 mount_helper_common.go:34] Warning: mount cleanup skipped because path does not exist: /var/lib/kubelet/pods/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0/volumes/kubernetes.io~local-volume/local-storage03-crc Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.508767 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.508978 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509052 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509107 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509159 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509235 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509326 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509380 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509430 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wps7x\" (UniqueName: \"kubernetes.io/projected/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-kube-api-access-wps7x\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509482 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509537 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509588 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.509643 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.510580 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage03-crc" (OuterVolumeSpecName: "glance") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "local-storage03-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.513950 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-config-data" (OuterVolumeSpecName: "config-data") pod "ecee33ed-a8b9-4c53-8e25-311fe5f9757d" (UID: "ecee33ed-a8b9-4c53-8e25-311fe5f9757d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.541131 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.542519 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.546962 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.589065 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-config-data" (OuterVolumeSpecName: "config-data") pod "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" (UID: "a9ac260d-3fb7-4391-b5c3-bb05d27b79e0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.611247 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecee33ed-a8b9-4c53-8e25-311fe5f9757d-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.611307 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.611321 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.611339 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.611390 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.627157 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage03-crc" (UniqueName: "kubernetes.io/local-volume/local-storage03-crc") on node "crc" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.712368 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5rggd\" (UniqueName: \"kubernetes.io/projected/0267ab5c-7b30-465a-89e3-67af886a3348-kube-api-access-5rggd\") pod \"0267ab5c-7b30-465a-89e3-67af886a3348\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.712904 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0267ab5c-7b30-465a-89e3-67af886a3348-operator-scripts\") pod \"0267ab5c-7b30-465a-89e3-67af886a3348\" (UID: \"0267ab5c-7b30-465a-89e3-67af886a3348\") " Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.714748 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage03-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage03-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.714314 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0267ab5c-7b30-465a-89e3-67af886a3348-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "0267ab5c-7b30-465a-89e3-67af886a3348" (UID: "0267ab5c-7b30-465a-89e3-67af886a3348"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.716443 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0267ab5c-7b30-465a-89e3-67af886a3348-kube-api-access-5rggd" (OuterVolumeSpecName: "kube-api-access-5rggd") pod "0267ab5c-7b30-465a-89e3-67af886a3348" (UID: "0267ab5c-7b30-465a-89e3-67af886a3348"). InnerVolumeSpecName "kube-api-access-5rggd". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.806843 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.814310 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-external-api-0"] Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.816958 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5rggd\" (UniqueName: \"kubernetes.io/projected/0267ab5c-7b30-465a-89e3-67af886a3348-kube-api-access-5rggd\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:13 crc kubenswrapper[4956]: I1126 17:16:13.816993 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/0267ab5c-7b30-465a-89e3-67af886a3348-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:14 crc kubenswrapper[4956]: I1126 17:16:14.174407 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" event={"ID":"0267ab5c-7b30-465a-89e3-67af886a3348","Type":"ContainerDied","Data":"188614107c454e3548120cd59345f0da9b835fba46179eaa93e20893450086d2"} Nov 26 17:16:14 crc kubenswrapper[4956]: I1126 17:16:14.174940 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="188614107c454e3548120cd59345f0da9b835fba46179eaa93e20893450086d2" Nov 26 17:16:14 crc kubenswrapper[4956]: I1126 17:16:14.174455 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-internal-api-0" Nov 26 17:16:14 crc kubenswrapper[4956]: I1126 17:16:14.174430 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glanced650-account-delete-bjkfl" Nov 26 17:16:14 crc kubenswrapper[4956]: I1126 17:16:14.220344 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:16:14 crc kubenswrapper[4956]: I1126 17:16:14.227806 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-internal-api-0"] Nov 26 17:16:15 crc kubenswrapper[4956]: I1126 17:16:15.004942 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" path="/var/lib/kubelet/pods/a9ac260d-3fb7-4391-b5c3-bb05d27b79e0/volumes" Nov 26 17:16:15 crc kubenswrapper[4956]: I1126 17:16:15.005821 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e164968b-dfde-4d68-b73e-632722876b76" path="/var/lib/kubelet/pods/e164968b-dfde-4d68-b73e-632722876b76/volumes" Nov 26 17:16:15 crc kubenswrapper[4956]: I1126 17:16:15.006749 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" path="/var/lib/kubelet/pods/ecee33ed-a8b9-4c53-8e25-311fe5f9757d/volumes" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.490014 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-4m8fj"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.503963 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-4m8fj"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.511590 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-d650-account-create-update-72bds"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.517459 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glanced650-account-delete-bjkfl"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.523225 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glanced650-account-delete-bjkfl"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.529093 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-d650-account-create-update-72bds"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580388 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-fmlrm"] Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.580818 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0267ab5c-7b30-465a-89e3-67af886a3348" containerName="mariadb-account-delete" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580849 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0267ab5c-7b30-465a-89e3-67af886a3348" containerName="mariadb-account-delete" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.580890 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d473f051-a35a-4c66-86ac-21c54348895e" containerName="openstackclient" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580901 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="d473f051-a35a-4c66-86ac-21c54348895e" containerName="openstackclient" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.580913 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580922 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.580934 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580942 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.580959 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580967 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.580981 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.580991 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.581004 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581016 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.581038 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581062 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.581076 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581085 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: E1126 17:16:19.581096 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581107 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581268 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="d473f051-a35a-4c66-86ac-21c54348895e" containerName="openstackclient" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581289 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581301 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581312 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="6e2d74c2-ac47-4731-a3e5-641274858dad" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581320 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="a9ac260d-3fb7-4391-b5c3-bb05d27b79e0" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581335 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581348 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="0267ab5c-7b30-465a-89e3-67af886a3348" containerName="mariadb-account-delete" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581356 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581367 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecee33ed-a8b9-4c53-8e25-311fe5f9757d" containerName="glance-httpd" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.581380 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="e164968b-dfde-4d68-b73e-632722876b76" containerName="glance-log" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.582080 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.588397 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-f567-account-create-update-j8wzt"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.589894 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.592278 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.594982 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fmlrm"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.605454 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f567-account-create-update-j8wzt"] Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.715046 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c82e33c-c999-4ac1-a9a9-770948eab705-operator-scripts\") pod \"glance-f567-account-create-update-j8wzt\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.715118 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2zkbr\" (UniqueName: \"kubernetes.io/projected/d275f55f-c0e8-47e2-bf81-1621f1d66a76-kube-api-access-2zkbr\") pod \"glance-db-create-fmlrm\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.715198 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d275f55f-c0e8-47e2-bf81-1621f1d66a76-operator-scripts\") pod \"glance-db-create-fmlrm\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.715229 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-98qdb\" (UniqueName: \"kubernetes.io/projected/2c82e33c-c999-4ac1-a9a9-770948eab705-kube-api-access-98qdb\") pod \"glance-f567-account-create-update-j8wzt\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.817197 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c82e33c-c999-4ac1-a9a9-770948eab705-operator-scripts\") pod \"glance-f567-account-create-update-j8wzt\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.817268 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2zkbr\" (UniqueName: \"kubernetes.io/projected/d275f55f-c0e8-47e2-bf81-1621f1d66a76-kube-api-access-2zkbr\") pod \"glance-db-create-fmlrm\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.817333 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d275f55f-c0e8-47e2-bf81-1621f1d66a76-operator-scripts\") pod \"glance-db-create-fmlrm\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.817362 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-98qdb\" (UniqueName: \"kubernetes.io/projected/2c82e33c-c999-4ac1-a9a9-770948eab705-kube-api-access-98qdb\") pod \"glance-f567-account-create-update-j8wzt\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.818257 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c82e33c-c999-4ac1-a9a9-770948eab705-operator-scripts\") pod \"glance-f567-account-create-update-j8wzt\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.818284 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d275f55f-c0e8-47e2-bf81-1621f1d66a76-operator-scripts\") pod \"glance-db-create-fmlrm\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.842717 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-98qdb\" (UniqueName: \"kubernetes.io/projected/2c82e33c-c999-4ac1-a9a9-770948eab705-kube-api-access-98qdb\") pod \"glance-f567-account-create-update-j8wzt\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.843887 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2zkbr\" (UniqueName: \"kubernetes.io/projected/d275f55f-c0e8-47e2-bf81-1621f1d66a76-kube-api-access-2zkbr\") pod \"glance-db-create-fmlrm\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.920072 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:19 crc kubenswrapper[4956]: I1126 17:16:19.932549 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:20 crc kubenswrapper[4956]: I1126 17:16:20.475730 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-fmlrm"] Nov 26 17:16:20 crc kubenswrapper[4956]: I1126 17:16:20.494927 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-f567-account-create-update-j8wzt"] Nov 26 17:16:20 crc kubenswrapper[4956]: W1126 17:16:20.503520 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2c82e33c_c999_4ac1_a9a9_770948eab705.slice/crio-7f6c53243326a9a9e2b2d77343a789ee1c762f86af5fc6429a24b41bcfef7731 WatchSource:0}: Error finding container 7f6c53243326a9a9e2b2d77343a789ee1c762f86af5fc6429a24b41bcfef7731: Status 404 returned error can't find the container with id 7f6c53243326a9a9e2b2d77343a789ee1c762f86af5fc6429a24b41bcfef7731 Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.004411 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0267ab5c-7b30-465a-89e3-67af886a3348" path="/var/lib/kubelet/pods/0267ab5c-7b30-465a-89e3-67af886a3348/volumes" Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.005497 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fcfb1fa-bda6-4346-ae62-73f9df0c69dc" path="/var/lib/kubelet/pods/4fcfb1fa-bda6-4346-ae62-73f9df0c69dc/volumes" Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.006074 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c1be27b4-e03f-4b93-9f1e-2eb79b980791" path="/var/lib/kubelet/pods/c1be27b4-e03f-4b93-9f1e-2eb79b980791/volumes" Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.240929 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" event={"ID":"2c82e33c-c999-4ac1-a9a9-770948eab705","Type":"ContainerStarted","Data":"a18e29b90b2f5c83e72cf855dd4f468da0b9e982cd4ed16c6a74511f7fa3e9a6"} Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.241010 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" event={"ID":"2c82e33c-c999-4ac1-a9a9-770948eab705","Type":"ContainerStarted","Data":"7f6c53243326a9a9e2b2d77343a789ee1c762f86af5fc6429a24b41bcfef7731"} Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.242459 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fmlrm" event={"ID":"d275f55f-c0e8-47e2-bf81-1621f1d66a76","Type":"ContainerStarted","Data":"ef8f20a63088a1211acf6f328879a9ee13c75df423b8d7b70dbe43264d39713b"} Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.242506 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fmlrm" event={"ID":"d275f55f-c0e8-47e2-bf81-1621f1d66a76","Type":"ContainerStarted","Data":"bb67e034273d49edcc3bda90e8c9f71f265a7fbdc82308ee17c2678dca1a2871"} Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.256792 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" podStartSLOduration=2.256763485 podStartE2EDuration="2.256763485s" podCreationTimestamp="2025-11-26 17:16:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:16:21.255276442 +0000 UTC m=+1426.951237004" watchObservedRunningTime="2025-11-26 17:16:21.256763485 +0000 UTC m=+1426.952724037" Nov 26 17:16:21 crc kubenswrapper[4956]: I1126 17:16:21.272640 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-create-fmlrm" podStartSLOduration=2.272617462 podStartE2EDuration="2.272617462s" podCreationTimestamp="2025-11-26 17:16:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:16:21.270394388 +0000 UTC m=+1426.966354940" watchObservedRunningTime="2025-11-26 17:16:21.272617462 +0000 UTC m=+1426.968578014" Nov 26 17:16:23 crc kubenswrapper[4956]: I1126 17:16:23.265070 4956 generic.go:334] "Generic (PLEG): container finished" podID="d275f55f-c0e8-47e2-bf81-1621f1d66a76" containerID="ef8f20a63088a1211acf6f328879a9ee13c75df423b8d7b70dbe43264d39713b" exitCode=0 Nov 26 17:16:23 crc kubenswrapper[4956]: I1126 17:16:23.265312 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fmlrm" event={"ID":"d275f55f-c0e8-47e2-bf81-1621f1d66a76","Type":"ContainerDied","Data":"ef8f20a63088a1211acf6f328879a9ee13c75df423b8d7b70dbe43264d39713b"} Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.606103 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.724290 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2zkbr\" (UniqueName: \"kubernetes.io/projected/d275f55f-c0e8-47e2-bf81-1621f1d66a76-kube-api-access-2zkbr\") pod \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.724649 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d275f55f-c0e8-47e2-bf81-1621f1d66a76-operator-scripts\") pod \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\" (UID: \"d275f55f-c0e8-47e2-bf81-1621f1d66a76\") " Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.725694 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d275f55f-c0e8-47e2-bf81-1621f1d66a76-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "d275f55f-c0e8-47e2-bf81-1621f1d66a76" (UID: "d275f55f-c0e8-47e2-bf81-1621f1d66a76"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.745838 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d275f55f-c0e8-47e2-bf81-1621f1d66a76-kube-api-access-2zkbr" (OuterVolumeSpecName: "kube-api-access-2zkbr") pod "d275f55f-c0e8-47e2-bf81-1621f1d66a76" (UID: "d275f55f-c0e8-47e2-bf81-1621f1d66a76"). InnerVolumeSpecName "kube-api-access-2zkbr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.833354 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2zkbr\" (UniqueName: \"kubernetes.io/projected/d275f55f-c0e8-47e2-bf81-1621f1d66a76-kube-api-access-2zkbr\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:24 crc kubenswrapper[4956]: I1126 17:16:24.833410 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/d275f55f-c0e8-47e2-bf81-1621f1d66a76-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:25 crc kubenswrapper[4956]: I1126 17:16:25.289367 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-fmlrm" event={"ID":"d275f55f-c0e8-47e2-bf81-1621f1d66a76","Type":"ContainerDied","Data":"bb67e034273d49edcc3bda90e8c9f71f265a7fbdc82308ee17c2678dca1a2871"} Nov 26 17:16:25 crc kubenswrapper[4956]: I1126 17:16:25.289901 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bb67e034273d49edcc3bda90e8c9f71f265a7fbdc82308ee17c2678dca1a2871" Nov 26 17:16:25 crc kubenswrapper[4956]: I1126 17:16:25.289395 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-fmlrm" Nov 26 17:16:25 crc kubenswrapper[4956]: I1126 17:16:25.291604 4956 generic.go:334] "Generic (PLEG): container finished" podID="2c82e33c-c999-4ac1-a9a9-770948eab705" containerID="a18e29b90b2f5c83e72cf855dd4f468da0b9e982cd4ed16c6a74511f7fa3e9a6" exitCode=0 Nov 26 17:16:25 crc kubenswrapper[4956]: I1126 17:16:25.291675 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" event={"ID":"2c82e33c-c999-4ac1-a9a9-770948eab705","Type":"ContainerDied","Data":"a18e29b90b2f5c83e72cf855dd4f468da0b9e982cd4ed16c6a74511f7fa3e9a6"} Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.627898 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.768966 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-98qdb\" (UniqueName: \"kubernetes.io/projected/2c82e33c-c999-4ac1-a9a9-770948eab705-kube-api-access-98qdb\") pod \"2c82e33c-c999-4ac1-a9a9-770948eab705\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.770210 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c82e33c-c999-4ac1-a9a9-770948eab705-operator-scripts\") pod \"2c82e33c-c999-4ac1-a9a9-770948eab705\" (UID: \"2c82e33c-c999-4ac1-a9a9-770948eab705\") " Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.771083 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c82e33c-c999-4ac1-a9a9-770948eab705-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "2c82e33c-c999-4ac1-a9a9-770948eab705" (UID: "2c82e33c-c999-4ac1-a9a9-770948eab705"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.775215 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c82e33c-c999-4ac1-a9a9-770948eab705-kube-api-access-98qdb" (OuterVolumeSpecName: "kube-api-access-98qdb") pod "2c82e33c-c999-4ac1-a9a9-770948eab705" (UID: "2c82e33c-c999-4ac1-a9a9-770948eab705"). InnerVolumeSpecName "kube-api-access-98qdb". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.873085 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-98qdb\" (UniqueName: \"kubernetes.io/projected/2c82e33c-c999-4ac1-a9a9-770948eab705-kube-api-access-98qdb\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:26 crc kubenswrapper[4956]: I1126 17:16:26.873165 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2c82e33c-c999-4ac1-a9a9-770948eab705-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:27 crc kubenswrapper[4956]: I1126 17:16:27.312991 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" event={"ID":"2c82e33c-c999-4ac1-a9a9-770948eab705","Type":"ContainerDied","Data":"7f6c53243326a9a9e2b2d77343a789ee1c762f86af5fc6429a24b41bcfef7731"} Nov 26 17:16:27 crc kubenswrapper[4956]: I1126 17:16:27.313572 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7f6c53243326a9a9e2b2d77343a789ee1c762f86af5fc6429a24b41bcfef7731" Nov 26 17:16:27 crc kubenswrapper[4956]: I1126 17:16:27.313640 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-f567-account-create-update-j8wzt" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.860030 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-hc2md"] Nov 26 17:16:29 crc kubenswrapper[4956]: E1126 17:16:29.860922 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d275f55f-c0e8-47e2-bf81-1621f1d66a76" containerName="mariadb-database-create" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.860941 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="d275f55f-c0e8-47e2-bf81-1621f1d66a76" containerName="mariadb-database-create" Nov 26 17:16:29 crc kubenswrapper[4956]: E1126 17:16:29.860968 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c82e33c-c999-4ac1-a9a9-770948eab705" containerName="mariadb-account-create-update" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.860975 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c82e33c-c999-4ac1-a9a9-770948eab705" containerName="mariadb-account-create-update" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.861111 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="d275f55f-c0e8-47e2-bf81-1621f1d66a76" containerName="mariadb-database-create" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.861128 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c82e33c-c999-4ac1-a9a9-770948eab705" containerName="mariadb-account-create-update" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.861720 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.864726 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qt6c9" Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.871280 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hc2md"] Nov 26 17:16:29 crc kubenswrapper[4956]: I1126 17:16:29.872729 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.037951 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5d6ch\" (UniqueName: \"kubernetes.io/projected/2fdba013-2bd9-4c44-be91-a2490b083e1c-kube-api-access-5d6ch\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.038041 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-db-sync-config-data\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.038075 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-config-data\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.140038 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5d6ch\" (UniqueName: \"kubernetes.io/projected/2fdba013-2bd9-4c44-be91-a2490b083e1c-kube-api-access-5d6ch\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.140564 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-db-sync-config-data\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.140596 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-config-data\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.148163 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-db-sync-config-data\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.148399 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-config-data\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.168776 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5d6ch\" (UniqueName: \"kubernetes.io/projected/2fdba013-2bd9-4c44-be91-a2490b083e1c-kube-api-access-5d6ch\") pod \"glance-db-sync-hc2md\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.194650 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.684338 4956 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.691305 4956 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.691816 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.692786 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2" gracePeriod=15 Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.692839 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a" gracePeriod=15 Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.693003 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227" gracePeriod=15 Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.693089 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba" gracePeriod=15 Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.693155 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65" gracePeriod=15 Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.693612 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hc2md"] Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.693689 4956 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694090 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694104 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694120 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694127 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694145 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694151 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694164 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694172 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694182 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694188 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694200 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694206 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.694223 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694230 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694360 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694371 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694380 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694388 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694396 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.694404 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Nov 26 17:16:30 crc kubenswrapper[4956]: E1126 17:16:30.782734 4956 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.850776 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.850848 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.850924 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.850949 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.851012 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.851042 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.851090 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.851149 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953044 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953111 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953166 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953197 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953250 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953300 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953325 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953346 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953439 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953489 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953514 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953541 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953567 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953598 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953624 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:30 crc kubenswrapper[4956]: I1126 17:16:30.953653 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.084157 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:31 crc kubenswrapper[4956]: W1126 17:16:31.116304 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-406391a0b1f9d7b47ed507d73bb3a5c5587fbd190abe5a522464768c0e14c01e WatchSource:0}: Error finding container 406391a0b1f9d7b47ed507d73bb3a5c5587fbd190abe5a522464768c0e14c01e: Status 404 returned error can't find the container with id 406391a0b1f9d7b47ed507d73bb3a5c5587fbd190abe5a522464768c0e14c01e Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.356939 4956 generic.go:334] "Generic (PLEG): container finished" podID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" containerID="d0abb6914ca35eea14580a7958053585360705a7c300d2afa7c1cc86251dfed2" exitCode=0 Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.357050 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d72606f-03e5-4a3a-a98e-05e56db0d2eb","Type":"ContainerDied","Data":"d0abb6914ca35eea14580a7958053585360705a7c300d2afa7c1cc86251dfed2"} Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.358430 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.359429 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hc2md" event={"ID":"2fdba013-2bd9-4c44-be91-a2490b083e1c","Type":"ContainerStarted","Data":"2c8863980fef4152580adb19b6d3d66bab0c60fe53cf432c863b14809e7bfa04"} Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.361078 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"406391a0b1f9d7b47ed507d73bb3a5c5587fbd190abe5a522464768c0e14c01e"} Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.364595 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.366043 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 17:16:31 crc kubenswrapper[4956]: I1126 17:16:31.367040 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65" exitCode=2 Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.381577 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42"} Nov 26 17:16:32 crc kubenswrapper[4956]: E1126 17:16:32.382377 4956 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.382525 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.384166 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.386150 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.386939 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a" exitCode=0 Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.386970 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227" exitCode=0 Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.386979 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba" exitCode=0 Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.387065 4956 scope.go:117] "RemoveContainer" containerID="a7b90cc372956d874f512b7968b8bb4b1c26a498309c0d50733eeeedd7e432b6" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.390568 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hc2md" event={"ID":"2fdba013-2bd9-4c44-be91-a2490b083e1c","Type":"ContainerStarted","Data":"5e56230a8c816380d0c0f85d11a40729b126bb9da2304acba577e7689001cf70"} Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.390681 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.390857 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.678135 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.678948 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.679514 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.785457 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kubelet-dir\") pod \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.786197 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-var-lock\") pod \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.785683 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "6d72606f-03e5-4a3a-a98e-05e56db0d2eb" (UID: "6d72606f-03e5-4a3a-a98e-05e56db0d2eb"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.786313 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-var-lock" (OuterVolumeSpecName: "var-lock") pod "6d72606f-03e5-4a3a-a98e-05e56db0d2eb" (UID: "6d72606f-03e5-4a3a-a98e-05e56db0d2eb"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.786351 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kube-api-access\") pod \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\" (UID: \"6d72606f-03e5-4a3a-a98e-05e56db0d2eb\") " Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.787393 4956 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kubelet-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.787418 4956 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-var-lock\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.816355 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "6d72606f-03e5-4a3a-a98e-05e56db0d2eb" (UID: "6d72606f-03e5-4a3a-a98e-05e56db0d2eb"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:32 crc kubenswrapper[4956]: I1126 17:16:32.888679 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/6d72606f-03e5-4a3a-a98e-05e56db0d2eb-kube-api-access\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.306584 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.307621 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.308328 4956 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.308578 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.308822 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.396327 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.396473 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.396538 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.396941 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.396972 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.396988 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.401336 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.401847 4956 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2" exitCode=0 Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.401980 4956 scope.go:117] "RemoveContainer" containerID="437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.402110 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.413629 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.413967 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"6d72606f-03e5-4a3a-a98e-05e56db0d2eb","Type":"ContainerDied","Data":"f990776e5ff12c2695a1a4ca7a44c87725040c74f072ffaab3b3c711d94c6ea1"} Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.414027 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="f990776e5ff12c2695a1a4ca7a44c87725040c74f072ffaab3b3c711d94c6ea1" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.416960 4956 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.428009 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.428625 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.428941 4956 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.429169 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.429352 4956 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.429532 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.442795 4956 scope.go:117] "RemoveContainer" containerID="8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.461408 4956 scope.go:117] "RemoveContainer" containerID="00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.481547 4956 scope.go:117] "RemoveContainer" containerID="0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.500433 4956 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.500466 4956 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.500477 4956 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.518319 4956 scope.go:117] "RemoveContainer" containerID="16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.563438 4956 scope.go:117] "RemoveContainer" containerID="8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.588088 4956 scope.go:117] "RemoveContainer" containerID="437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.588781 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\": container with ID starting with 437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a not found: ID does not exist" containerID="437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.588827 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a"} err="failed to get container status \"437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\": rpc error: code = NotFound desc = could not find container \"437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a\": container with ID starting with 437b4db9dad9af7d78a4c17ae1bf32393d44b49416e244ac3864ad8867b9ea2a not found: ID does not exist" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.588857 4956 scope.go:117] "RemoveContainer" containerID="8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.589106 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\": container with ID starting with 8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227 not found: ID does not exist" containerID="8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.589130 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227"} err="failed to get container status \"8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\": rpc error: code = NotFound desc = could not find container \"8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227\": container with ID starting with 8a2df8b43cf4142b5e83feffadfadde217e7ac7db7238c5082786cef74e20227 not found: ID does not exist" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.589144 4956 scope.go:117] "RemoveContainer" containerID="00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.589438 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\": container with ID starting with 00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba not found: ID does not exist" containerID="00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.589500 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba"} err="failed to get container status \"00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\": rpc error: code = NotFound desc = could not find container \"00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba\": container with ID starting with 00d58e703c29df152c60f494e6562e760304f306b09638f224d789a1967351ba not found: ID does not exist" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.589549 4956 scope.go:117] "RemoveContainer" containerID="0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.589916 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\": container with ID starting with 0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65 not found: ID does not exist" containerID="0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.589986 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65"} err="failed to get container status \"0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\": rpc error: code = NotFound desc = could not find container \"0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65\": container with ID starting with 0227b08ab4ea274cacece9f9954d37911cd8cb601c4667a47e6ba3faecc7be65 not found: ID does not exist" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.590011 4956 scope.go:117] "RemoveContainer" containerID="16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.590402 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\": container with ID starting with 16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2 not found: ID does not exist" containerID="16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.590437 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2"} err="failed to get container status \"16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\": rpc error: code = NotFound desc = could not find container \"16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2\": container with ID starting with 16dd94153d20ceda0116f8ba8744f2df9f0549336fdd765310ccfa0bbd28aef2 not found: ID does not exist" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.590459 4956 scope.go:117] "RemoveContainer" containerID="8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18" Nov 26 17:16:33 crc kubenswrapper[4956]: E1126 17:16:33.590841 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\": container with ID starting with 8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18 not found: ID does not exist" containerID="8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18" Nov 26 17:16:33 crc kubenswrapper[4956]: I1126 17:16:33.590888 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18"} err="failed to get container status \"8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\": rpc error: code = NotFound desc = could not find container \"8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18\": container with ID starting with 8f751f061a3e408e3b1d6f81e3b5e101c549d805400b22aefa7d19abd9609a18 not found: ID does not exist" Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.000842 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.002111 4956 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.002728 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.006886 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.437341 4956 generic.go:334] "Generic (PLEG): container finished" podID="2fdba013-2bd9-4c44-be91-a2490b083e1c" containerID="5e56230a8c816380d0c0f85d11a40729b126bb9da2304acba577e7689001cf70" exitCode=0 Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.437462 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hc2md" event={"ID":"2fdba013-2bd9-4c44-be91-a2490b083e1c","Type":"ContainerDied","Data":"5e56230a8c816380d0c0f85d11a40729b126bb9da2304acba577e7689001cf70"} Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.438997 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:35 crc kubenswrapper[4956]: I1126 17:16:35.439678 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:35 crc kubenswrapper[4956]: E1126 17:16:35.735384 4956 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-crc.187b9dfd65095bdc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Killing,Message:Stopping container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 17:16:30.693080028 +0000 UTC m=+1436.389040590,LastTimestamp:2025-11-26 17:16:30.693080028 +0000 UTC m=+1436.389040590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 17:16:36 crc kubenswrapper[4956]: E1126 17:16:36.339989 4956 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.9:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-crc.187b9dfd65095bdc openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Killing,Message:Stopping container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-11-26 17:16:30.693080028 +0000 UTC m=+1436.389040590,LastTimestamp:2025-11-26 17:16:30.693080028 +0000 UTC m=+1436.389040590,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.742442 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.743536 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.743956 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.859663 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-db-sync-config-data\") pod \"2fdba013-2bd9-4c44-be91-a2490b083e1c\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.859743 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-config-data\") pod \"2fdba013-2bd9-4c44-be91-a2490b083e1c\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.859857 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5d6ch\" (UniqueName: \"kubernetes.io/projected/2fdba013-2bd9-4c44-be91-a2490b083e1c-kube-api-access-5d6ch\") pod \"2fdba013-2bd9-4c44-be91-a2490b083e1c\" (UID: \"2fdba013-2bd9-4c44-be91-a2490b083e1c\") " Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.865644 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fdba013-2bd9-4c44-be91-a2490b083e1c-kube-api-access-5d6ch" (OuterVolumeSpecName: "kube-api-access-5d6ch") pod "2fdba013-2bd9-4c44-be91-a2490b083e1c" (UID: "2fdba013-2bd9-4c44-be91-a2490b083e1c"). InnerVolumeSpecName "kube-api-access-5d6ch". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.869333 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "2fdba013-2bd9-4c44-be91-a2490b083e1c" (UID: "2fdba013-2bd9-4c44-be91-a2490b083e1c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.900379 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-config-data" (OuterVolumeSpecName: "config-data") pod "2fdba013-2bd9-4c44-be91-a2490b083e1c" (UID: "2fdba013-2bd9-4c44-be91-a2490b083e1c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.961263 4956 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.961295 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2fdba013-2bd9-4c44-be91-a2490b083e1c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:36 crc kubenswrapper[4956]: I1126 17:16:36.961305 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5d6ch\" (UniqueName: \"kubernetes.io/projected/2fdba013-2bd9-4c44-be91-a2490b083e1c-kube-api-access-5d6ch\") on node \"crc\" DevicePath \"\"" Nov 26 17:16:37 crc kubenswrapper[4956]: I1126 17:16:37.469086 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-hc2md" event={"ID":"2fdba013-2bd9-4c44-be91-a2490b083e1c","Type":"ContainerDied","Data":"2c8863980fef4152580adb19b6d3d66bab0c60fe53cf432c863b14809e7bfa04"} Nov 26 17:16:37 crc kubenswrapper[4956]: I1126 17:16:37.469610 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2c8863980fef4152580adb19b6d3d66bab0c60fe53cf432c863b14809e7bfa04" Nov 26 17:16:37 crc kubenswrapper[4956]: I1126 17:16:37.469215 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-hc2md" Nov 26 17:16:37 crc kubenswrapper[4956]: I1126 17:16:37.475152 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:37 crc kubenswrapper[4956]: I1126 17:16:37.475903 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.575756 4956 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.577321 4956 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.577897 4956 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.578272 4956 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.579047 4956 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:39 crc kubenswrapper[4956]: I1126 17:16:39.579095 4956 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.579555 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="200ms" Nov 26 17:16:39 crc kubenswrapper[4956]: E1126 17:16:39.781415 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="400ms" Nov 26 17:16:40 crc kubenswrapper[4956]: E1126 17:16:40.183272 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="800ms" Nov 26 17:16:40 crc kubenswrapper[4956]: E1126 17:16:40.984213 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="1.6s" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.512346 4956 generic.go:334] "Generic (PLEG): container finished" podID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" containerID="c5eaa44ed20bf03ec8473cda3e7348f1ca69b6e79c0fc8b463d29691c017b92a" exitCode=1 Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.512472 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" event={"ID":"1090fc6e-22d8-4fb2-9296-46edde4e1af5","Type":"ContainerDied","Data":"c5eaa44ed20bf03ec8473cda3e7348f1ca69b6e79c0fc8b463d29691c017b92a"} Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.513247 4956 scope.go:117] "RemoveContainer" containerID="c5eaa44ed20bf03ec8473cda3e7348f1ca69b6e79c0fc8b463d29691c017b92a" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.513847 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.515170 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.515651 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.516457 4956 generic.go:334] "Generic (PLEG): container finished" podID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" containerID="fa4d811b64502f501dbebbb3bd74b87ace4ad7364aa8b664e8656f59082be9b1" exitCode=1 Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.516515 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" event={"ID":"91276233-6cc3-408e-b9e4-e8a4c7ecc60d","Type":"ContainerDied","Data":"fa4d811b64502f501dbebbb3bd74b87ace4ad7364aa8b664e8656f59082be9b1"} Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.517586 4956 scope.go:117] "RemoveContainer" containerID="fa4d811b64502f501dbebbb3bd74b87ace4ad7364aa8b664e8656f59082be9b1" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.519703 4956 generic.go:334] "Generic (PLEG): container finished" podID="7f723082-bb1f-4dd1-beb2-c1553728eba8" containerID="3a76257a664da6b026934dc765eb9fb3c30538c411787a826217022c03664890" exitCode=1 Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.519740 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" event={"ID":"7f723082-bb1f-4dd1-beb2-c1553728eba8","Type":"ContainerDied","Data":"3a76257a664da6b026934dc765eb9fb3c30538c411787a826217022c03664890"} Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.520567 4956 scope.go:117] "RemoveContainer" containerID="3a76257a664da6b026934dc765eb9fb3c30538c411787a826217022c03664890" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.520764 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.521512 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.522339 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.522617 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.523162 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.523587 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.524095 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.524463 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:41 crc kubenswrapper[4956]: I1126 17:16:41.524817 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.146339 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.146746 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.537310 4956 generic.go:334] "Generic (PLEG): container finished" podID="7f723082-bb1f-4dd1-beb2-c1553728eba8" containerID="bf69bb93a90993a446adb0947a5a117fb3c453de67975e299c7800968af6965d" exitCode=1 Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.537397 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" event={"ID":"7f723082-bb1f-4dd1-beb2-c1553728eba8","Type":"ContainerDied","Data":"bf69bb93a90993a446adb0947a5a117fb3c453de67975e299c7800968af6965d"} Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.537442 4956 scope.go:117] "RemoveContainer" containerID="3a76257a664da6b026934dc765eb9fb3c30538c411787a826217022c03664890" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.537983 4956 scope.go:117] "RemoveContainer" containerID="bf69bb93a90993a446adb0947a5a117fb3c453de67975e299c7800968af6965d" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.538255 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: E1126 17:16:42.538332 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-69b487756b-xzvnf_openstack-operators(7f723082-bb1f-4dd1-beb2-c1553728eba8)\"" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.538475 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.539346 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.540159 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.540486 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.542378 4956 generic.go:334] "Generic (PLEG): container finished" podID="d2238f24-43e6-4512-880a-1c3b10ba3844" containerID="90329b5d7c5468486a729f757cfd442e663d9c08c31c3bb62235099f0dfcf5c5" exitCode=1 Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.542494 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" event={"ID":"d2238f24-43e6-4512-880a-1c3b10ba3844","Type":"ContainerDied","Data":"90329b5d7c5468486a729f757cfd442e663d9c08c31c3bb62235099f0dfcf5c5"} Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.543279 4956 scope.go:117] "RemoveContainer" containerID="90329b5d7c5468486a729f757cfd442e663d9c08c31c3bb62235099f0dfcf5c5" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.546205 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.546596 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.546814 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.547229 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.547884 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.548155 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.549841 4956 generic.go:334] "Generic (PLEG): container finished" podID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" containerID="1ad575fea9a0131098b5c443e9cedf4290f62b9bed95bf742b26201366167ec7" exitCode=1 Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.549944 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" event={"ID":"1090fc6e-22d8-4fb2-9296-46edde4e1af5","Type":"ContainerDied","Data":"1ad575fea9a0131098b5c443e9cedf4290f62b9bed95bf742b26201366167ec7"} Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.550460 4956 scope.go:117] "RemoveContainer" containerID="1ad575fea9a0131098b5c443e9cedf4290f62b9bed95bf742b26201366167ec7" Nov 26 17:16:42 crc kubenswrapper[4956]: E1126 17:16:42.550755 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-f659d5cfc-2plpb_openstack-operators(1090fc6e-22d8-4fb2-9296-46edde4e1af5)\"" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.550972 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.551801 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.552278 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.552538 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.552896 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.553286 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.554130 4956 generic.go:334] "Generic (PLEG): container finished" podID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" containerID="dfb90541317792b4217df3fad3be8ded2c8630b4550de8f6e64c20b9024f753e" exitCode=1 Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.554213 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" event={"ID":"25d0fbbb-c310-46d0-ae91-a9804761e8a3","Type":"ContainerDied","Data":"dfb90541317792b4217df3fad3be8ded2c8630b4550de8f6e64c20b9024f753e"} Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.554982 4956 scope.go:117] "RemoveContainer" containerID="dfb90541317792b4217df3fad3be8ded2c8630b4550de8f6e64c20b9024f753e" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.555141 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.555588 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.555832 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.556212 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.556697 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.557068 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.557394 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.558785 4956 generic.go:334] "Generic (PLEG): container finished" podID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" containerID="82c8f9fae8432e92d62df2370fe79333d1d191a56c3baf7c6d61f5c236d1d57f" exitCode=1 Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.558856 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" event={"ID":"91276233-6cc3-408e-b9e4-e8a4c7ecc60d","Type":"ContainerDied","Data":"82c8f9fae8432e92d62df2370fe79333d1d191a56c3baf7c6d61f5c236d1d57f"} Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.559384 4956 scope.go:117] "RemoveContainer" containerID="82c8f9fae8432e92d62df2370fe79333d1d191a56c3baf7c6d61f5c236d1d57f" Nov 26 17:16:42 crc kubenswrapper[4956]: E1126 17:16:42.559700 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-7674ccc9cf-j5f9c_openstack-operators(91276233-6cc3-408e-b9e4-e8a4c7ecc60d)\"" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.560189 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.560547 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.560974 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.561454 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.561697 4956 generic.go:334] "Generic (PLEG): container finished" podID="063794be-3564-44a8-8613-a966e3ebd953" containerID="9ed1d2310d68aeae6b94618f4b7e1f11aff32d6885c3549ffaddd6bc9b04e23a" exitCode=1 Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.561747 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerDied","Data":"9ed1d2310d68aeae6b94618f4b7e1f11aff32d6885c3549ffaddd6bc9b04e23a"} Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.562566 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.562721 4956 scope.go:117] "RemoveContainer" containerID="9ed1d2310d68aeae6b94618f4b7e1f11aff32d6885c3549ffaddd6bc9b04e23a" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.563018 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.563364 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.564120 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.564696 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.565229 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.565655 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.566087 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.566491 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.570383 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.570798 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: E1126 17:16:42.586662 4956 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.9:6443: connect: connection refused" interval="3.2s" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.690554 4956 scope.go:117] "RemoveContainer" containerID="c5eaa44ed20bf03ec8473cda3e7348f1ca69b6e79c0fc8b463d29691c017b92a" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.814515 4956 scope.go:117] "RemoveContainer" containerID="fa4d811b64502f501dbebbb3bd74b87ace4ad7364aa8b664e8656f59082be9b1" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.995210 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.996777 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.997249 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.997598 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.997896 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.998143 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.998498 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.998668 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:42 crc kubenswrapper[4956]: I1126 17:16:42.998966 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.082746 4956 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.082816 4956 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:43 crc kubenswrapper[4956]: E1126 17:16:43.083403 4956 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.084094 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:43 crc kubenswrapper[4956]: W1126 17:16:43.115976 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod71bb4a3aecc4ba5b26c4b7318770ce13.slice/crio-d730f1f6d7bd0afc8edbd8eaba68155777a8394ba226f43475f5207b8cb287c3 WatchSource:0}: Error finding container d730f1f6d7bd0afc8edbd8eaba68155777a8394ba226f43475f5207b8cb287c3: Status 404 returned error can't find the container with id d730f1f6d7bd0afc8edbd8eaba68155777a8394ba226f43475f5207b8cb287c3 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.572916 4956 generic.go:334] "Generic (PLEG): container finished" podID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" containerID="1d9619c912d81fb74629584afb7da1b67bdb423d9058bcea0044bb6daacb5035" exitCode=1 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.573024 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" event={"ID":"25d0fbbb-c310-46d0-ae91-a9804761e8a3","Type":"ContainerDied","Data":"1d9619c912d81fb74629584afb7da1b67bdb423d9058bcea0044bb6daacb5035"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.573073 4956 scope.go:117] "RemoveContainer" containerID="dfb90541317792b4217df3fad3be8ded2c8630b4550de8f6e64c20b9024f753e" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.574112 4956 scope.go:117] "RemoveContainer" containerID="1d9619c912d81fb74629584afb7da1b67bdb423d9058bcea0044bb6daacb5035" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.574214 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.574482 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: E1126 17:16:43.574597 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 10s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-94bjc_openstack-operators(25d0fbbb-c310-46d0-ae91-a9804761e8a3)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.574713 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.575053 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.575585 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.575938 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.576266 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.576525 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.576949 4956 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="3a1179613c34fd59dc940dff3ef6ed02b4933ab9f1371100e797e97d45a7a9b1" exitCode=0 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.577036 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"3a1179613c34fd59dc940dff3ef6ed02b4933ab9f1371100e797e97d45a7a9b1"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.577080 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d730f1f6d7bd0afc8edbd8eaba68155777a8394ba226f43475f5207b8cb287c3"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.577456 4956 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.577480 4956 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:43 crc kubenswrapper[4956]: E1126 17:16:43.577778 4956 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.578830 4956 generic.go:334] "Generic (PLEG): container finished" podID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" containerID="c9a78b3017292779ce3a1dab640bfee03067a0aa57e9353b9942d952daa07a7a" exitCode=1 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.578884 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" event={"ID":"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5","Type":"ContainerDied","Data":"c9a78b3017292779ce3a1dab640bfee03067a0aa57e9353b9942d952daa07a7a"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.580448 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.580595 4956 scope.go:117] "RemoveContainer" containerID="c9a78b3017292779ce3a1dab640bfee03067a0aa57e9353b9942d952daa07a7a" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.580728 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.581004 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.581276 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.581611 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.581751 4956 generic.go:334] "Generic (PLEG): container finished" podID="063794be-3564-44a8-8613-a966e3ebd953" containerID="c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65" exitCode=1 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.581817 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerDied","Data":"c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.581920 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.582207 4956 scope.go:117] "RemoveContainer" containerID="c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.582220 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: E1126 17:16:43.582405 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-5c4ffc67cb-dnf99_openstack-operators(063794be-3564-44a8-8613-a966e3ebd953)\"" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podUID="063794be-3564-44a8-8613-a966e3ebd953" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.582675 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.583541 4956 status_manager.go:851] "Failed to get status for pod" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d556748c5-l2bl4\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.583697 4956 generic.go:334] "Generic (PLEG): container finished" podID="5adb08fd-ab2a-4d2d-afab-4b605419b741" containerID="2c3e3e9bc5551b703165baa12edde7bfcf326212884f441003436889505bb0a2" exitCode=1 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.583758 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" event={"ID":"5adb08fd-ab2a-4d2d-afab-4b605419b741","Type":"ContainerDied","Data":"2c3e3e9bc5551b703165baa12edde7bfcf326212884f441003436889505bb0a2"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.584091 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.584122 4956 scope.go:117] "RemoveContainer" containerID="2c3e3e9bc5551b703165baa12edde7bfcf326212884f441003436889505bb0a2" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.584542 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.585413 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.586190 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.586960 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.587407 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.587973 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.588419 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.588904 4956 status_manager.go:851] "Failed to get status for pod" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-dd7556475-2wzb2\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.589197 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.589502 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.589823 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.590182 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.590480 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.590718 4956 status_manager.go:851] "Failed to get status for pod" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d556748c5-l2bl4\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.591078 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.591532 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.592145 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.593228 4956 scope.go:117] "RemoveContainer" containerID="bf69bb93a90993a446adb0947a5a117fb3c453de67975e299c7800968af6965d" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.593269 4956 status_manager.go:851] "Failed to get status for pod" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d556748c5-l2bl4\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: E1126 17:16:43.593998 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=keystone-operator-controller-manager-69b487756b-xzvnf_openstack-operators(7f723082-bb1f-4dd1-beb2-c1553728eba8)\"" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.594100 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.594679 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.594912 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.595074 4956 status_manager.go:851] "Failed to get status for pod" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-dd7556475-2wzb2\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.595238 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.595538 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.595965 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.596312 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.596749 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.597815 4956 generic.go:334] "Generic (PLEG): container finished" podID="d2238f24-43e6-4512-880a-1c3b10ba3844" containerID="68a20f81b082aa1a0c8823900ad871b8b28d7144bb3e288ba3de998b2a55a7ed" exitCode=1 Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.597878 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" event={"ID":"d2238f24-43e6-4512-880a-1c3b10ba3844","Type":"ContainerDied","Data":"68a20f81b082aa1a0c8823900ad871b8b28d7144bb3e288ba3de998b2a55a7ed"} Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.598378 4956 scope.go:117] "RemoveContainer" containerID="68a20f81b082aa1a0c8823900ad871b8b28d7144bb3e288ba3de998b2a55a7ed" Nov 26 17:16:43 crc kubenswrapper[4956]: E1126 17:16:43.598627 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-788c9777f6-7qmf6_openstack-operators(d2238f24-43e6-4512-880a-1c3b10ba3844)\"" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.598902 4956 status_manager.go:851] "Failed to get status for pod" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/horizon-operator-controller-manager-dd7556475-2wzb2\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.599253 4956 status_manager.go:851] "Failed to get status for pod" podUID="063794be-3564-44a8-8613-a966e3ebd953" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/infra-operator-controller-manager-5c4ffc67cb-dnf99\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.600238 4956 status_manager.go:851] "Failed to get status for pod" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/mariadb-operator-controller-manager-7674ccc9cf-j5f9c\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.601406 4956 status_manager.go:851] "Failed to get status for pod" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/swift-operator-controller-manager-f659d5cfc-2plpb\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.602632 4956 status_manager.go:851] "Failed to get status for pod" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" pod="glance-kuttl-tests/glance-db-sync-hc2md" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/glance-kuttl-tests/pods/glance-db-sync-hc2md\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.602968 4956 status_manager.go:851] "Failed to get status for pod" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/rabbitmq-cluster-operator-779fc9694b-94bjc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.603158 4956 status_manager.go:851] "Failed to get status for pod" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/metallb-system/pods/metallb-operator-controller-manager-6d556748c5-l2bl4\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.603552 4956 status_manager.go:851] "Failed to get status for pod" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/glance-operator-controller-manager-788c9777f6-7qmf6\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.603778 4956 status_manager.go:851] "Failed to get status for pod" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.604213 4956 status_manager.go:851] "Failed to get status for pod" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openstack-operators/pods/keystone-operator-controller-manager-69b487756b-xzvnf\": dial tcp 38.102.83.9:6443: connect: connection refused" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.724663 4956 scope.go:117] "RemoveContainer" containerID="9ed1d2310d68aeae6b94618f4b7e1f11aff32d6885c3549ffaddd6bc9b04e23a" Nov 26 17:16:43 crc kubenswrapper[4956]: I1126 17:16:43.845030 4956 scope.go:117] "RemoveContainer" containerID="90329b5d7c5468486a729f757cfd442e663d9c08c31c3bb62235099f0dfcf5c5" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.075137 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.496992 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.497475 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.498522 4956 scope.go:117] "RemoveContainer" containerID="82c8f9fae8432e92d62df2370fe79333d1d191a56c3baf7c6d61f5c236d1d57f" Nov 26 17:16:44 crc kubenswrapper[4956]: E1126 17:16:44.498841 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=mariadb-operator-controller-manager-7674ccc9cf-j5f9c_openstack-operators(91276233-6cc3-408e-b9e4-e8a4c7ecc60d)\"" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.625201 4956 generic.go:334] "Generic (PLEG): container finished" podID="5adb08fd-ab2a-4d2d-afab-4b605419b741" containerID="91da8e2ac76c3df09848b0fb434c251f530215420da0d257aad64bed62022e41" exitCode=1 Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.625241 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" event={"ID":"5adb08fd-ab2a-4d2d-afab-4b605419b741","Type":"ContainerDied","Data":"91da8e2ac76c3df09848b0fb434c251f530215420da0d257aad64bed62022e41"} Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.625379 4956 scope.go:117] "RemoveContainer" containerID="2c3e3e9bc5551b703165baa12edde7bfcf326212884f441003436889505bb0a2" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.626089 4956 scope.go:117] "RemoveContainer" containerID="91da8e2ac76c3df09848b0fb434c251f530215420da0d257aad64bed62022e41" Nov 26 17:16:44 crc kubenswrapper[4956]: E1126 17:16:44.626340 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-dd7556475-2wzb2_openstack-operators(5adb08fd-ab2a-4d2d-afab-4b605419b741)\"" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.633700 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"6aa21f98cf9f8afcde3865b479dccfdd188411a894ff159ef2d9e85c93e75876"} Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.633761 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d2cc145b43ea19389177a90b2b8f03c1748b8e114b1f4b8eec3010bb83e12f72"} Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.637390 4956 generic.go:334] "Generic (PLEG): container finished" podID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" containerID="db86fb47da84f7cf27d06410e683cbb750bd7d88e7ef6e14a2ebcfc52da5cad1" exitCode=1 Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.637446 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" event={"ID":"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5","Type":"ContainerDied","Data":"db86fb47da84f7cf27d06410e683cbb750bd7d88e7ef6e14a2ebcfc52da5cad1"} Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.639519 4956 scope.go:117] "RemoveContainer" containerID="db86fb47da84f7cf27d06410e683cbb750bd7d88e7ef6e14a2ebcfc52da5cad1" Nov 26 17:16:44 crc kubenswrapper[4956]: E1126 17:16:44.639938 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-6d556748c5-l2bl4_metallb-system(4b5ac5a5-46f2-4fc2-944d-58fcafb997b5)\"" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" Nov 26 17:16:44 crc kubenswrapper[4956]: I1126 17:16:44.682663 4956 scope.go:117] "RemoveContainer" containerID="c9a78b3017292779ce3a1dab640bfee03067a0aa57e9353b9942d952daa07a7a" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.649616 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.650054 4956 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d" exitCode=1 Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.650177 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d"} Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.650774 4956 scope.go:117] "RemoveContainer" containerID="a0a39fe27a525c817ba18541482278dd10dd7ddbc47d9c89b119a1640c26039d" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.658138 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"2c5259ad02ecfec778a0090dc0cc3045502134777fb45d12ad402d7e0bde6ba9"} Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.658196 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3dde6f2b51dad789e1d14698f3f5a950de5c23c165bf8d9215e4462e42fc157c"} Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.658214 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"3ab9aca694635018c0df36bc05efb08b6464dd01955b02d2dccae7b29fe2d22b"} Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.658338 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.658403 4956 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.658427 4956 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.661110 4956 scope.go:117] "RemoveContainer" containerID="db86fb47da84f7cf27d06410e683cbb750bd7d88e7ef6e14a2ebcfc52da5cad1" Nov 26 17:16:45 crc kubenswrapper[4956]: E1126 17:16:45.661358 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=metallb-operator-controller-manager-6d556748c5-l2bl4_metallb-system(4b5ac5a5-46f2-4fc2-944d-58fcafb997b5)\"" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.684426 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.685155 4956 scope.go:117] "RemoveContainer" containerID="c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.685308 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:16:45 crc kubenswrapper[4956]: E1126 17:16:45.685526 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-5c4ffc67cb-dnf99_openstack-operators(063794be-3564-44a8-8613-a966e3ebd953)\"" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podUID="063794be-3564-44a8-8613-a966e3ebd953" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.845900 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:16:45 crc kubenswrapper[4956]: I1126 17:16:45.846284 4956 scope.go:117] "RemoveContainer" containerID="68a20f81b082aa1a0c8823900ad871b8b28d7144bb3e288ba3de998b2a55a7ed" Nov 26 17:16:45 crc kubenswrapper[4956]: E1126 17:16:45.846480 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=glance-operator-controller-manager-788c9777f6-7qmf6_openstack-operators(d2238f24-43e6-4512-880a-1c3b10ba3844)\"" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" Nov 26 17:16:46 crc kubenswrapper[4956]: I1126 17:16:46.673037 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Nov 26 17:16:46 crc kubenswrapper[4956]: I1126 17:16:46.673115 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"1cf5e57f75050c5e355be51147e96674c33ee85397a058ada04c27dfa2250a35"} Nov 26 17:16:46 crc kubenswrapper[4956]: I1126 17:16:46.673779 4956 scope.go:117] "RemoveContainer" containerID="c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65" Nov 26 17:16:46 crc kubenswrapper[4956]: E1126 17:16:46.691068 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=infra-operator-controller-manager-5c4ffc67cb-dnf99_openstack-operators(063794be-3564-44a8-8613-a966e3ebd953)\"" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podUID="063794be-3564-44a8-8613-a966e3ebd953" Nov 26 17:16:46 crc kubenswrapper[4956]: I1126 17:16:46.699083 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 17:16:46 crc kubenswrapper[4956]: I1126 17:16:46.704132 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 17:16:47 crc kubenswrapper[4956]: I1126 17:16:47.681584 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 17:16:48 crc kubenswrapper[4956]: I1126 17:16:48.085155 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:48 crc kubenswrapper[4956]: I1126 17:16:48.085546 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:48 crc kubenswrapper[4956]: I1126 17:16:48.092815 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:48 crc kubenswrapper[4956]: I1126 17:16:48.426066 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:16:48 crc kubenswrapper[4956]: I1126 17:16:48.426136 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:16:48 crc kubenswrapper[4956]: I1126 17:16:48.426715 4956 scope.go:117] "RemoveContainer" containerID="1ad575fea9a0131098b5c443e9cedf4290f62b9bed95bf742b26201366167ec7" Nov 26 17:16:48 crc kubenswrapper[4956]: E1126 17:16:48.427009 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=swift-operator-controller-manager-f659d5cfc-2plpb_openstack-operators(1090fc6e-22d8-4fb2-9296-46edde4e1af5)\"" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" Nov 26 17:16:49 crc kubenswrapper[4956]: I1126 17:16:49.615487 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:16:49 crc kubenswrapper[4956]: I1126 17:16:49.616587 4956 scope.go:117] "RemoveContainer" containerID="91da8e2ac76c3df09848b0fb434c251f530215420da0d257aad64bed62022e41" Nov 26 17:16:49 crc kubenswrapper[4956]: E1126 17:16:49.616859 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 10s restarting failed container=manager pod=horizon-operator-controller-manager-dd7556475-2wzb2_openstack-operators(5adb08fd-ab2a-4d2d-afab-4b605419b741)\"" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" Nov 26 17:16:50 crc kubenswrapper[4956]: I1126 17:16:50.681174 4956 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:50 crc kubenswrapper[4956]: I1126 17:16:50.909924 4956 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6ac1faeb-b0c7-4f26-b733-a45b8317b768" Nov 26 17:16:51 crc kubenswrapper[4956]: I1126 17:16:51.718353 4956 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:51 crc kubenswrapper[4956]: I1126 17:16:51.718397 4956 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:51 crc kubenswrapper[4956]: I1126 17:16:51.724144 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:16:51 crc kubenswrapper[4956]: I1126 17:16:51.724630 4956 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6ac1faeb-b0c7-4f26-b733-a45b8317b768" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.147205 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.148083 4956 scope.go:117] "RemoveContainer" containerID="bf69bb93a90993a446adb0947a5a117fb3c453de67975e299c7800968af6965d" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.729608 4956 generic.go:334] "Generic (PLEG): container finished" podID="7f723082-bb1f-4dd1-beb2-c1553728eba8" containerID="1b44fb416299bc7baf820c4a234582a47fffb8dbfa7bae05b8818b19e05e2d9f" exitCode=1 Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.729698 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" event={"ID":"7f723082-bb1f-4dd1-beb2-c1553728eba8","Type":"ContainerDied","Data":"1b44fb416299bc7baf820c4a234582a47fffb8dbfa7bae05b8818b19e05e2d9f"} Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.730195 4956 scope.go:117] "RemoveContainer" containerID="bf69bb93a90993a446adb0947a5a117fb3c453de67975e299c7800968af6965d" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.730478 4956 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.730529 4956 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="1c5fd9ff-5bc2-49ef-a042-8042c97467ee" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.731340 4956 scope.go:117] "RemoveContainer" containerID="1b44fb416299bc7baf820c4a234582a47fffb8dbfa7bae05b8818b19e05e2d9f" Nov 26 17:16:52 crc kubenswrapper[4956]: E1126 17:16:52.731778 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-69b487756b-xzvnf_openstack-operators(7f723082-bb1f-4dd1-beb2-c1553728eba8)\"" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" Nov 26 17:16:52 crc kubenswrapper[4956]: I1126 17:16:52.757842 4956 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="6ac1faeb-b0c7-4f26-b733-a45b8317b768" Nov 26 17:16:54 crc kubenswrapper[4956]: I1126 17:16:54.075686 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:16:54 crc kubenswrapper[4956]: I1126 17:16:54.077668 4956 scope.go:117] "RemoveContainer" containerID="db86fb47da84f7cf27d06410e683cbb750bd7d88e7ef6e14a2ebcfc52da5cad1" Nov 26 17:16:54 crc kubenswrapper[4956]: I1126 17:16:54.751983 4956 generic.go:334] "Generic (PLEG): container finished" podID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" containerID="96790e14b14035abb0c672bfd89eb397bce2832cdaa10f5c84b866bb6cb83c7b" exitCode=1 Nov 26 17:16:54 crc kubenswrapper[4956]: I1126 17:16:54.752075 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" event={"ID":"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5","Type":"ContainerDied","Data":"96790e14b14035abb0c672bfd89eb397bce2832cdaa10f5c84b866bb6cb83c7b"} Nov 26 17:16:54 crc kubenswrapper[4956]: I1126 17:16:54.752446 4956 scope.go:117] "RemoveContainer" containerID="db86fb47da84f7cf27d06410e683cbb750bd7d88e7ef6e14a2ebcfc52da5cad1" Nov 26 17:16:54 crc kubenswrapper[4956]: I1126 17:16:54.753775 4956 scope.go:117] "RemoveContainer" containerID="96790e14b14035abb0c672bfd89eb397bce2832cdaa10f5c84b866bb6cb83c7b" Nov 26 17:16:54 crc kubenswrapper[4956]: E1126 17:16:54.754064 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-6d556748c5-l2bl4_metallb-system(4b5ac5a5-46f2-4fc2-944d-58fcafb997b5)\"" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.008818 4956 scope.go:117] "RemoveContainer" containerID="82c8f9fae8432e92d62df2370fe79333d1d191a56c3baf7c6d61f5c236d1d57f" Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.762703 4956 generic.go:334] "Generic (PLEG): container finished" podID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" containerID="260bd047d307fe123919de5cf89208c3c6deb1835cf7959c708183b015ada582" exitCode=1 Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.762790 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" event={"ID":"91276233-6cc3-408e-b9e4-e8a4c7ecc60d","Type":"ContainerDied","Data":"260bd047d307fe123919de5cf89208c3c6deb1835cf7959c708183b015ada582"} Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.762895 4956 scope.go:117] "RemoveContainer" containerID="82c8f9fae8432e92d62df2370fe79333d1d191a56c3baf7c6d61f5c236d1d57f" Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.763667 4956 scope.go:117] "RemoveContainer" containerID="260bd047d307fe123919de5cf89208c3c6deb1835cf7959c708183b015ada582" Nov 26 17:16:55 crc kubenswrapper[4956]: E1126 17:16:55.764717 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-7674ccc9cf-j5f9c_openstack-operators(91276233-6cc3-408e-b9e4-e8a4c7ecc60d)\"" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.812575 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.813568 4956 scope.go:117] "RemoveContainer" containerID="68a20f81b082aa1a0c8823900ad871b8b28d7144bb3e288ba3de998b2a55a7ed" Nov 26 17:16:55 crc kubenswrapper[4956]: I1126 17:16:55.996977 4956 scope.go:117] "RemoveContainer" containerID="1d9619c912d81fb74629584afb7da1b67bdb423d9058bcea0044bb6daacb5035" Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.779125 4956 generic.go:334] "Generic (PLEG): container finished" podID="d2238f24-43e6-4512-880a-1c3b10ba3844" containerID="7394812381c2102727732d462b7add087310ffb9df19d4ba13e9c5901ddfa5bb" exitCode=1 Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.779172 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" event={"ID":"d2238f24-43e6-4512-880a-1c3b10ba3844","Type":"ContainerDied","Data":"7394812381c2102727732d462b7add087310ffb9df19d4ba13e9c5901ddfa5bb"} Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.779221 4956 scope.go:117] "RemoveContainer" containerID="68a20f81b082aa1a0c8823900ad871b8b28d7144bb3e288ba3de998b2a55a7ed" Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.779839 4956 scope.go:117] "RemoveContainer" containerID="7394812381c2102727732d462b7add087310ffb9df19d4ba13e9c5901ddfa5bb" Nov 26 17:16:56 crc kubenswrapper[4956]: E1126 17:16:56.780112 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-788c9777f6-7qmf6_openstack-operators(d2238f24-43e6-4512-880a-1c3b10ba3844)\"" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.783062 4956 generic.go:334] "Generic (PLEG): container finished" podID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" containerID="2749dd177cffbcfc2dbc8b7955c07550d2fcfd512092421305688ec0003b48a5" exitCode=1 Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.783111 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" event={"ID":"25d0fbbb-c310-46d0-ae91-a9804761e8a3","Type":"ContainerDied","Data":"2749dd177cffbcfc2dbc8b7955c07550d2fcfd512092421305688ec0003b48a5"} Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.783749 4956 scope.go:117] "RemoveContainer" containerID="2749dd177cffbcfc2dbc8b7955c07550d2fcfd512092421305688ec0003b48a5" Nov 26 17:16:56 crc kubenswrapper[4956]: E1126 17:16:56.784032 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-94bjc_openstack-operators(25d0fbbb-c310-46d0-ae91-a9804761e8a3)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" Nov 26 17:16:56 crc kubenswrapper[4956]: I1126 17:16:56.851516 4956 scope.go:117] "RemoveContainer" containerID="1d9619c912d81fb74629584afb7da1b67bdb423d9058bcea0044bb6daacb5035" Nov 26 17:16:59 crc kubenswrapper[4956]: I1126 17:16:59.615060 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:16:59 crc kubenswrapper[4956]: I1126 17:16:59.617328 4956 scope.go:117] "RemoveContainer" containerID="91da8e2ac76c3df09848b0fb434c251f530215420da0d257aad64bed62022e41" Nov 26 17:17:00 crc kubenswrapper[4956]: I1126 17:17:00.093198 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Nov 26 17:17:00 crc kubenswrapper[4956]: I1126 17:17:00.822658 4956 generic.go:334] "Generic (PLEG): container finished" podID="5adb08fd-ab2a-4d2d-afab-4b605419b741" containerID="bc6765a30483f9a81087c59ad95b0d7ca9e3dd3edf016fe4d18e4920e137bc78" exitCode=1 Nov 26 17:17:00 crc kubenswrapper[4956]: I1126 17:17:00.823239 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" event={"ID":"5adb08fd-ab2a-4d2d-afab-4b605419b741","Type":"ContainerDied","Data":"bc6765a30483f9a81087c59ad95b0d7ca9e3dd3edf016fe4d18e4920e137bc78"} Nov 26 17:17:00 crc kubenswrapper[4956]: I1126 17:17:00.823296 4956 scope.go:117] "RemoveContainer" containerID="91da8e2ac76c3df09848b0fb434c251f530215420da0d257aad64bed62022e41" Nov 26 17:17:00 crc kubenswrapper[4956]: I1126 17:17:00.824246 4956 scope.go:117] "RemoveContainer" containerID="bc6765a30483f9a81087c59ad95b0d7ca9e3dd3edf016fe4d18e4920e137bc78" Nov 26 17:17:00 crc kubenswrapper[4956]: E1126 17:17:00.824546 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-dd7556475-2wzb2_openstack-operators(5adb08fd-ab2a-4d2d-afab-4b605419b741)\"" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" Nov 26 17:17:01 crc kubenswrapper[4956]: I1126 17:17:01.122758 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Nov 26 17:17:01 crc kubenswrapper[4956]: I1126 17:17:01.378302 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Nov 26 17:17:01 crc kubenswrapper[4956]: I1126 17:17:01.622585 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-server-dockercfg-cjjk9" Nov 26 17:17:01 crc kubenswrapper[4956]: I1126 17:17:01.814391 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Nov 26 17:17:01 crc kubenswrapper[4956]: I1126 17:17:01.903020 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Nov 26 17:17:01 crc kubenswrapper[4956]: I1126 17:17:01.995681 4956 scope.go:117] "RemoveContainer" containerID="c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.147120 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.147654 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.148481 4956 scope.go:117] "RemoveContainer" containerID="1b44fb416299bc7baf820c4a234582a47fffb8dbfa7bae05b8818b19e05e2d9f" Nov 26 17:17:02 crc kubenswrapper[4956]: E1126 17:17:02.148726 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=keystone-operator-controller-manager-69b487756b-xzvnf_openstack-operators(7f723082-bb1f-4dd1-beb2-c1553728eba8)\"" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" podUID="7f723082-bb1f-4dd1-beb2-c1553728eba8" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.180456 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.242075 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.294252 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.374300 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.459277 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-lvvl8" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.646232 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.749969 4956 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.847703 4956 generic.go:334] "Generic (PLEG): container finished" podID="063794be-3564-44a8-8613-a966e3ebd953" containerID="0bb8080a76c32a1016437944da9e1814ea71a4f651db57b69b8ee5e5be90991e" exitCode=1 Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.847733 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerDied","Data":"0bb8080a76c32a1016437944da9e1814ea71a4f651db57b69b8ee5e5be90991e"} Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.848679 4956 scope.go:117] "RemoveContainer" containerID="c56b4cacebd9da1f5cf0e195129d9006b1d6b5afaae993a477a0881d8fec6b65" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.849355 4956 scope.go:117] "RemoveContainer" containerID="0bb8080a76c32a1016437944da9e1814ea71a4f651db57b69b8ee5e5be90991e" Nov 26 17:17:02 crc kubenswrapper[4956]: E1126 17:17:02.849618 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-5c4ffc67cb-dnf99_openstack-operators(063794be-3564-44a8-8613-a966e3ebd953)\"" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podUID="063794be-3564-44a8-8613-a966e3ebd953" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.970277 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Nov 26 17:17:02 crc kubenswrapper[4956]: I1126 17:17:02.995877 4956 scope.go:117] "RemoveContainer" containerID="1ad575fea9a0131098b5c443e9cedf4290f62b9bed95bf742b26201366167ec7" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.358833 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.372673 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-storage-config-data" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.442854 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-default-user" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.466176 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.637516 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.712487 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-zs4z6" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.774367 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.846156 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.859752 4956 generic.go:334] "Generic (PLEG): container finished" podID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" containerID="a204b7151062c5f0f785290cca3cffaa99922125ee7984b9f2e8ab4dd65a806c" exitCode=1 Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.859821 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" event={"ID":"1090fc6e-22d8-4fb2-9296-46edde4e1af5","Type":"ContainerDied","Data":"a204b7151062c5f0f785290cca3cffaa99922125ee7984b9f2e8ab4dd65a806c"} Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.859899 4956 scope.go:117] "RemoveContainer" containerID="1ad575fea9a0131098b5c443e9cedf4290f62b9bed95bf742b26201366167ec7" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.861118 4956 scope.go:117] "RemoveContainer" containerID="a204b7151062c5f0f785290cca3cffaa99922125ee7984b9f2e8ab4dd65a806c" Nov 26 17:17:03 crc kubenswrapper[4956]: E1126 17:17:03.861993 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-f659d5cfc-2plpb_openstack-operators(1090fc6e-22d8-4fb2-9296-46edde4e1af5)\"" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.885616 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.887737 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.945785 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.963626 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Nov 26 17:17:03 crc kubenswrapper[4956]: I1126 17:17:03.978522 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.011047 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.075383 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.076185 4956 scope.go:117] "RemoveContainer" containerID="96790e14b14035abb0c672bfd89eb397bce2832cdaa10f5c84b866bb6cb83c7b" Nov 26 17:17:04 crc kubenswrapper[4956]: E1126 17:17:04.076455 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=metallb-operator-controller-manager-6d556748c5-l2bl4_metallb-system(4b5ac5a5-46f2-4fc2-944d-58fcafb997b5)\"" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" podUID="4b5ac5a5-46f2-4fc2-944d-58fcafb997b5" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.088104 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.291953 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.314214 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.463787 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.493620 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.497151 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.497238 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.498405 4956 scope.go:117] "RemoveContainer" containerID="260bd047d307fe123919de5cf89208c3c6deb1835cf7959c708183b015ada582" Nov 26 17:17:04 crc kubenswrapper[4956]: E1126 17:17:04.499130 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=mariadb-operator-controller-manager-7674ccc9cf-j5f9c_openstack-operators(91276233-6cc3-408e-b9e4-e8a4c7ecc60d)\"" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" podUID="91276233-6cc3-408e-b9e4-e8a4c7ecc60d" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.567416 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-index-dockercfg-p4h5f" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.575154 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.596893 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.696082 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-rr2js" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.765598 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.819783 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.831725 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.872553 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Nov 26 17:17:04 crc kubenswrapper[4956]: I1126 17:17:04.943054 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.312998 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.315089 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.395222 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.397023 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.406483 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.416819 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.436357 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.457339 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.502026 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.658711 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.684510 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.684594 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.685288 4956 scope.go:117] "RemoveContainer" containerID="0bb8080a76c32a1016437944da9e1814ea71a4f651db57b69b8ee5e5be90991e" Nov 26 17:17:05 crc kubenswrapper[4956]: E1126 17:17:05.685523 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-5c4ffc67cb-dnf99_openstack-operators(063794be-3564-44a8-8613-a966e3ebd953)\"" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podUID="063794be-3564-44a8-8613-a966e3ebd953" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.710389 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.766670 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.812693 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.813554 4956 scope.go:117] "RemoveContainer" containerID="7394812381c2102727732d462b7add087310ffb9df19d4ba13e9c5901ddfa5bb" Nov 26 17:17:05 crc kubenswrapper[4956]: E1126 17:17:05.813842 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-788c9777f6-7qmf6_openstack-operators(d2238f24-43e6-4512-880a-1c3b10ba3844)\"" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.849067 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.849165 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.921084 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.945540 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.963737 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Nov 26 17:17:05 crc kubenswrapper[4956]: I1126 17:17:05.991100 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.040351 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.095377 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-keystone-dockercfg-w6kjs" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.128293 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.138625 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-plugins-conf" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.216272 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.218515 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"memcached-config-data" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.312341 4956 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.328831 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.328938 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.346444 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.358854 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=16.358833398 podStartE2EDuration="16.358833398s" podCreationTimestamp="2025-11-26 17:16:50 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:17:06.35474535 +0000 UTC m=+1472.050705922" watchObservedRunningTime="2025-11-26 17:17:06.358833398 +0000 UTC m=+1472.054793950" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.408205 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.438193 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.471019 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-t7s7x" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.475257 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.518339 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.518364 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.519177 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.576686 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.581834 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.747174 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.791790 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.805410 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.822987 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.844675 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.861130 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.877897 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.911082 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Nov 26 17:17:06 crc kubenswrapper[4956]: I1126 17:17:06.913830 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.005105 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-lvklx" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.085591 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.087704 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.156828 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.173046 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.245080 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.437197 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.465299 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.472578 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.528723 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.549036 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.611733 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.667683 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.667882 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.719789 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"galera-openstack-dockercfg-f55gs" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.728222 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.750013 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.755901 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.757883 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.789303 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.791034 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.812596 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.834234 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.884066 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Nov 26 17:17:07 crc kubenswrapper[4956]: I1126 17:17:07.973706 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.034725 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.122144 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.127445 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-rwdd4" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.145991 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.154165 4956 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.222021 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-scripts" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.248347 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.256698 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.277163 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.305729 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.410120 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.411003 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.426126 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.426175 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.426691 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.427051 4956 scope.go:117] "RemoveContainer" containerID="a204b7151062c5f0f785290cca3cffaa99922125ee7984b9f2e8ab4dd65a806c" Nov 26 17:17:08 crc kubenswrapper[4956]: E1126 17:17:08.427374 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-f659d5cfc-2plpb_openstack-operators(1090fc6e-22d8-4fb2-9296-46edde4e1af5)\"" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.472564 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.486840 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.509921 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.526400 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.533030 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.656851 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.769786 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.769991 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.775248 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.849571 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.892243 4956 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Nov 26 17:17:08 crc kubenswrapper[4956]: I1126 17:17:08.992499 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.115526 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.124157 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.163264 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.169314 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"swift-ring-files" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.187507 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.188327 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-service-cert" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.216079 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.306118 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.327297 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.352090 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.381390 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.392613 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.417781 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openshift-service-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.510251 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.522341 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.552303 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.563133 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.599262 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.612667 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.614687 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.615525 4956 scope.go:117] "RemoveContainer" containerID="bc6765a30483f9a81087c59ad95b0d7ca9e3dd3edf016fe4d18e4920e137bc78" Nov 26 17:17:09 crc kubenswrapper[4956]: E1126 17:17:09.615752 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-dd7556475-2wzb2_openstack-operators(5adb08fd-ab2a-4d2d-afab-4b605419b741)\"" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.675980 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.689006 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.706955 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"memcached-memcached-dockercfg-8t9gl" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.718974 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.719370 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.790059 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.800405 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.822465 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Nov 26 17:17:09 crc kubenswrapper[4956]: I1126 17:17:09.992143 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.058692 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.076586 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.110550 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.121692 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.157889 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.179222 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.183853 4956 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.201611 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.334391 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"rabbitmq-server-conf" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.335049 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.395717 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-dockercfg-rlp4r" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.406206 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.526958 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.576378 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.704697 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.714775 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.789836 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.906938 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.947300 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.973064 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Nov 26 17:17:10 crc kubenswrapper[4956]: I1126 17:17:10.995484 4956 scope.go:117] "RemoveContainer" containerID="2749dd177cffbcfc2dbc8b7955c07550d2fcfd512092421305688ec0003b48a5" Nov 26 17:17:10 crc kubenswrapper[4956]: E1126 17:17:10.996004 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with CrashLoopBackOff: \"back-off 20s restarting failed container=operator pod=rabbitmq-cluster-operator-779fc9694b-94bjc_openstack-operators(25d0fbbb-c310-46d0-ae91-a9804761e8a3)\"" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" podUID="25d0fbbb-c310-46d0-ae91-a9804761e8a3" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.027571 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.067140 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.229608 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.289995 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.392291 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.470833 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.498170 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.498703 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.511650 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.579822 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.592459 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.594014 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.617678 4956 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.636308 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-index-dockercfg-j8svs" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.652497 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.673434 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.703945 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.706266 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.840013 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.864435 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.909062 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-index-dockercfg-xfgqg" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.911850 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.952911 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.961016 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Nov 26 17:17:11 crc kubenswrapper[4956]: I1126 17:17:11.979328 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.038138 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.045159 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.109327 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.111356 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.165254 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.198523 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.258296 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-index-dockercfg-6f5m5" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.310216 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-service-cert" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.340541 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.374457 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.376322 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.419279 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.487011 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.496104 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.517457 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.517858 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.553100 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.580654 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.662510 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.859694 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Nov 26 17:17:12 crc kubenswrapper[4956]: I1126 17:17:12.870474 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.021299 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.133795 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.186718 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.188512 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.253578 4956 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.253918 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42" gracePeriod=5 Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.262187 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.336975 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.339753 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-service-cert" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.353908 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-index-dockercfg-2htkq" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.364164 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.386579 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-proxy-config-data" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.390640 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"kube-root-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.439495 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.453149 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.498589 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.528758 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.659680 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.880440 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-index-dockercfg-l8tt5" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.914126 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.914577 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.925403 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Nov 26 17:17:13 crc kubenswrapper[4956]: I1126 17:17:13.937555 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"rabbitmq-erlang-cookie" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.008834 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.115359 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.157562 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-swift-dockercfg-zvkwr" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.204926 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.225241 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.253742 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.368651 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.392372 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.454051 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.472314 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.519233 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.695245 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.704347 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Nov 26 17:17:14 crc kubenswrapper[4956]: I1126 17:17:14.767966 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.046498 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.063003 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.128958 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.243585 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.284085 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.351857 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.382699 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.514889 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.528790 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.612044 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.647029 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.688674 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-service-cert" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.737729 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.775178 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config-data" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.812428 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.813207 4956 scope.go:117] "RemoveContainer" containerID="7394812381c2102727732d462b7add087310ffb9df19d4ba13e9c5901ddfa5bb" Nov 26 17:17:15 crc kubenswrapper[4956]: E1126 17:17:15.813557 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=glance-operator-controller-manager-788c9777f6-7qmf6_openstack-operators(d2238f24-43e6-4512-880a-1c3b10ba3844)\"" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" podUID="d2238f24-43e6-4512-880a-1c3b10ba3844" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.837420 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.949632 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-ldx5h" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.997656 4956 scope.go:117] "RemoveContainer" containerID="96790e14b14035abb0c672bfd89eb397bce2832cdaa10f5c84b866bb6cb83c7b" Nov 26 17:17:15 crc kubenswrapper[4956]: I1126 17:17:15.998007 4956 scope.go:117] "RemoveContainer" containerID="0bb8080a76c32a1016437944da9e1814ea71a4f651db57b69b8ee5e5be90991e" Nov 26 17:17:16 crc kubenswrapper[4956]: E1126 17:17:16.001007 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=infra-operator-controller-manager-5c4ffc67cb-dnf99_openstack-operators(063794be-3564-44a8-8613-a966e3ebd953)\"" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" podUID="063794be-3564-44a8-8613-a966e3ebd953" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.021996 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.029745 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.065207 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"keystone-config-data" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.249340 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.342245 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.369264 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-6jzcl" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.381054 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-service-cert" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.471190 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.484695 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.527174 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.607527 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.731827 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.799995 4956 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.870787 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.909138 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"swift-conf" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.969551 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.996447 4956 scope.go:117] "RemoveContainer" containerID="260bd047d307fe123919de5cf89208c3c6deb1835cf7959c708183b015ada582" Nov 26 17:17:16 crc kubenswrapper[4956]: I1126 17:17:16.997937 4956 scope.go:117] "RemoveContainer" containerID="1b44fb416299bc7baf820c4a234582a47fffb8dbfa7bae05b8818b19e05e2d9f" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.009408 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" event={"ID":"4b5ac5a5-46f2-4fc2-944d-58fcafb997b5","Type":"ContainerStarted","Data":"ca13bb3a42857879f1b5048b2ea35c43f826cba8fb28e993e472db5063113224"} Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.010816 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.106005 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.138677 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.240431 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.243997 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-tl4dj" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.430640 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-cjzx8" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.575190 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.626636 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.800698 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Nov 26 17:17:17 crc kubenswrapper[4956]: I1126 17:17:17.987273 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.020698 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" event={"ID":"7f723082-bb1f-4dd1-beb2-c1553728eba8","Type":"ContainerStarted","Data":"40a67662071de3d9bfc10659f8c0690dc625196f3d2da7ed54da5966b5372228"} Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.021495 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.024701 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" event={"ID":"91276233-6cc3-408e-b9e4-e8a4c7ecc60d","Type":"ContainerStarted","Data":"0358a1827ae38f86823ead8f6a634f5e3c9f378e5384d384eee2d7de1177c3a8"} Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.025202 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.188254 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.290539 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.325048 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-service-cert" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.459457 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-index-dockercfg-s25tb" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.521708 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.956167 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 26 17:17:18 crc kubenswrapper[4956]: I1126 17:17:18.956281 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.035452 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.035516 4956 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42" exitCode=137 Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.035603 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.035602 4956 scope.go:117] "RemoveContainer" containerID="a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.062385 4956 scope.go:117] "RemoveContainer" containerID="a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42" Nov 26 17:17:19 crc kubenswrapper[4956]: E1126 17:17:19.063012 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42\": container with ID starting with a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42 not found: ID does not exist" containerID="a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.063054 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42"} err="failed to get container status \"a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42\": rpc error: code = NotFound desc = could not find container \"a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42\": container with ID starting with a6e916ea47132d978e08347a1cfe654f833dc2b2304b6345ade80c3b65ca2d42 not found: ID does not exist" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086683 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086795 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086848 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086854 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086887 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086948 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.086977 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.087037 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.087115 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.087708 4956 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.087727 4956 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.087736 4956 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.087747 4956 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.097222 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.112012 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.189379 4956 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.274569 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.614933 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:17:19 crc kubenswrapper[4956]: I1126 17:17:19.615915 4956 scope.go:117] "RemoveContainer" containerID="bc6765a30483f9a81087c59ad95b0d7ca9e3dd3edf016fe4d18e4920e137bc78" Nov 26 17:17:19 crc kubenswrapper[4956]: E1126 17:17:19.616325 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=horizon-operator-controller-manager-dd7556475-2wzb2_openstack-operators(5adb08fd-ab2a-4d2d-afab-4b605419b741)\"" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" podUID="5adb08fd-ab2a-4d2d-afab-4b605419b741" Nov 26 17:17:20 crc kubenswrapper[4956]: I1126 17:17:20.996122 4956 scope.go:117] "RemoveContainer" containerID="a204b7151062c5f0f785290cca3cffaa99922125ee7984b9f2e8ab4dd65a806c" Nov 26 17:17:20 crc kubenswrapper[4956]: E1126 17:17:20.996385 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with CrashLoopBackOff: \"back-off 20s restarting failed container=manager pod=swift-operator-controller-manager-f659d5cfc-2plpb_openstack-operators(1090fc6e-22d8-4fb2-9296-46edde4e1af5)\"" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" podUID="1090fc6e-22d8-4fb2-9296-46edde4e1af5" Nov 26 17:17:21 crc kubenswrapper[4956]: I1126 17:17:21.004096 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Nov 26 17:17:22 crc kubenswrapper[4956]: I1126 17:17:22.152294 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-69b487756b-xzvnf" Nov 26 17:17:23 crc kubenswrapper[4956]: I1126 17:17:23.995452 4956 scope.go:117] "RemoveContainer" containerID="2749dd177cffbcfc2dbc8b7955c07550d2fcfd512092421305688ec0003b48a5" Nov 26 17:17:24 crc kubenswrapper[4956]: I1126 17:17:24.501674 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-7674ccc9cf-j5f9c" Nov 26 17:17:25 crc kubenswrapper[4956]: I1126 17:17:25.105644 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-779fc9694b-94bjc" event={"ID":"25d0fbbb-c310-46d0-ae91-a9804761e8a3","Type":"ContainerStarted","Data":"205cfb3829dd63503d019578fcabba215da971c6ff6a704b71d98381e6105bb8"} Nov 26 17:17:27 crc kubenswrapper[4956]: I1126 17:17:27.995832 4956 scope.go:117] "RemoveContainer" containerID="7394812381c2102727732d462b7add087310ffb9df19d4ba13e9c5901ddfa5bb" Nov 26 17:17:29 crc kubenswrapper[4956]: I1126 17:17:29.141563 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" event={"ID":"d2238f24-43e6-4512-880a-1c3b10ba3844","Type":"ContainerStarted","Data":"bf7ea078ed44ef602e9a4fde16673014bd82aa44c2e8c3e05a808851dfa3169b"} Nov 26 17:17:29 crc kubenswrapper[4956]: I1126 17:17:29.142249 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:17:30 crc kubenswrapper[4956]: I1126 17:17:30.996404 4956 scope.go:117] "RemoveContainer" containerID="0bb8080a76c32a1016437944da9e1814ea71a4f651db57b69b8ee5e5be90991e" Nov 26 17:17:32 crc kubenswrapper[4956]: I1126 17:17:32.177787 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" event={"ID":"063794be-3564-44a8-8613-a966e3ebd953","Type":"ContainerStarted","Data":"42cae39888bbb87be1bf5ad1cd5a73ac6ffe6115766f30fc874d26afdddb3d22"} Nov 26 17:17:32 crc kubenswrapper[4956]: I1126 17:17:32.178903 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:17:34 crc kubenswrapper[4956]: I1126 17:17:34.998914 4956 scope.go:117] "RemoveContainer" containerID="bc6765a30483f9a81087c59ad95b0d7ca9e3dd3edf016fe4d18e4920e137bc78" Nov 26 17:17:35 crc kubenswrapper[4956]: I1126 17:17:35.816705 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-788c9777f6-7qmf6" Nov 26 17:17:35 crc kubenswrapper[4956]: I1126 17:17:35.995918 4956 scope.go:117] "RemoveContainer" containerID="a204b7151062c5f0f785290cca3cffaa99922125ee7984b9f2e8ab4dd65a806c" Nov 26 17:17:36 crc kubenswrapper[4956]: I1126 17:17:36.212647 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" event={"ID":"5adb08fd-ab2a-4d2d-afab-4b605419b741","Type":"ContainerStarted","Data":"72eaa1ff19408c3c8333efd8a27c685f61483451f7960e24a81f292dd50c3c80"} Nov 26 17:17:36 crc kubenswrapper[4956]: I1126 17:17:36.213009 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:17:36 crc kubenswrapper[4956]: I1126 17:17:36.217548 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" event={"ID":"1090fc6e-22d8-4fb2-9296-46edde4e1af5","Type":"ContainerStarted","Data":"78697c1b971800b454eaf26f09b652a3924cfd35e531a5f5dafa1db4d2009416"} Nov 26 17:17:36 crc kubenswrapper[4956]: I1126 17:17:36.217809 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:17:39 crc kubenswrapper[4956]: I1126 17:17:39.551821 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:17:39 crc kubenswrapper[4956]: I1126 17:17:39.552326 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:17:40 crc kubenswrapper[4956]: I1126 17:17:40.134850 4956 scope.go:117] "RemoveContainer" containerID="03efd86766c47d8d2c6d010d2cb83100c25f1eb5e24996df860ea40aad397ec6" Nov 26 17:17:40 crc kubenswrapper[4956]: I1126 17:17:40.158115 4956 scope.go:117] "RemoveContainer" containerID="697de2465d9a9a21cb91c05eeda2c427180481c27c0e0f519e2a2949e4141996" Nov 26 17:17:40 crc kubenswrapper[4956]: I1126 17:17:40.189922 4956 scope.go:117] "RemoveContainer" containerID="33989093ae93333ad65ceb20f287cc009cdef1a1dfd04c6469371e1745fec63c" Nov 26 17:17:45 crc kubenswrapper[4956]: I1126 17:17:45.691240 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-5c4ffc67cb-dnf99" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.924857 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:46 crc kubenswrapper[4956]: E1126 17:17:46.925614 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" containerName="glance-db-sync" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.925630 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" containerName="glance-db-sync" Nov 26 17:17:46 crc kubenswrapper[4956]: E1126 17:17:46.925643 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.925649 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 17:17:46 crc kubenswrapper[4956]: E1126 17:17:46.925686 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" containerName="installer" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.925692 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" containerName="installer" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.925839 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.925851 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" containerName="glance-db-sync" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.925891 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="6d72606f-03e5-4a3a-a98e-05e56db0d2eb" containerName="installer" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.926641 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.932296 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.934485 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.934621 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qt6c9" Nov 26 17:17:46 crc kubenswrapper[4956]: I1126 17:17:46.993485 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112441 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-scripts\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112518 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-httpd-run\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112731 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112769 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112795 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g62zp\" (UniqueName: \"kubernetes.io/projected/30d410b5-3f0a-4528-ba02-9c89753e575c-kube-api-access-g62zp\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112919 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-dev\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112960 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.112991 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-config-data\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.113016 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-sys\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.113044 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.113080 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.113103 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-run\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.113156 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-logs\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.113882 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-lib-modules\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215725 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-dev\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215787 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215818 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-config-data\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215843 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-sys\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215880 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215919 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215938 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-run\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215971 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-logs\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.215969 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-dev\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216075 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-lib-modules\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216006 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-lib-modules\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216258 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-scripts\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216307 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216327 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-httpd-run\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216362 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216405 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g62zp\" (UniqueName: \"kubernetes.io/projected/30d410b5-3f0a-4528-ba02-9c89753e575c-kube-api-access-g62zp\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216519 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216606 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-run\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.216908 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.217136 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-logs\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.217181 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-sys\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.217228 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.217420 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-nvme\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.217467 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.218074 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-httpd-run\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.224922 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-scripts\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.224952 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-config-data\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.241743 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g62zp\" (UniqueName: \"kubernetes.io/projected/30d410b5-3f0a-4528-ba02-9c89753e575c-kube-api-access-g62zp\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.242288 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.242904 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.263844 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:47 crc kubenswrapper[4956]: I1126 17:17:47.776040 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:47 crc kubenswrapper[4956]: W1126 17:17:47.796429 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod30d410b5_3f0a_4528_ba02_9c89753e575c.slice/crio-dad5d86daeb2a1eaa12102047fb38c27cdc13bc585247159a6fdef3c9a1207bf WatchSource:0}: Error finding container dad5d86daeb2a1eaa12102047fb38c27cdc13bc585247159a6fdef3c9a1207bf: Status 404 returned error can't find the container with id dad5d86daeb2a1eaa12102047fb38c27cdc13bc585247159a6fdef3c9a1207bf Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.038005 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.325171 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"30d410b5-3f0a-4528-ba02-9c89753e575c","Type":"ContainerStarted","Data":"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc"} Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.325558 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-log" containerID="cri-o://6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5" gracePeriod=30 Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.325993 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-httpd" containerID="cri-o://1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc" gracePeriod=30 Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.326225 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"30d410b5-3f0a-4528-ba02-9c89753e575c","Type":"ContainerStarted","Data":"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5"} Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.326286 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"30d410b5-3f0a-4528-ba02-9c89753e575c","Type":"ContainerStarted","Data":"dad5d86daeb2a1eaa12102047fb38c27cdc13bc585247159a6fdef3c9a1207bf"} Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.358580 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.358557336 podStartE2EDuration="2.358557336s" podCreationTimestamp="2025-11-26 17:17:46 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:17:48.356395284 +0000 UTC m=+1514.052355846" watchObservedRunningTime="2025-11-26 17:17:48.358557336 +0000 UTC m=+1514.054517888" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.432038 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-f659d5cfc-2plpb" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.850396 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947500 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-logs\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947602 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947636 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-httpd-run\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947662 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-lib-modules\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947695 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-scripts\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947772 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-iscsi\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947829 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947858 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-dev\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947909 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-run\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947938 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-logs" (OuterVolumeSpecName: "logs") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947953 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-nvme\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.947994 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g62zp\" (UniqueName: \"kubernetes.io/projected/30d410b5-3f0a-4528-ba02-9c89753e575c-kube-api-access-g62zp\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948059 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-var-locks-brick\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948095 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-config-data\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948124 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-sys\") pod \"30d410b5-3f0a-4528-ba02-9c89753e575c\" (UID: \"30d410b5-3f0a-4528-ba02-9c89753e575c\") " Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948526 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948541 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-run" (OuterVolumeSpecName: "run") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948582 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-sys" (OuterVolumeSpecName: "sys") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948615 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-dev" (OuterVolumeSpecName: "dev") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948619 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948660 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948744 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.948813 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.949143 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.955443 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.957839 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-scripts" (OuterVolumeSpecName: "scripts") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.957941 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.964271 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/30d410b5-3f0a-4528-ba02-9c89753e575c-kube-api-access-g62zp" (OuterVolumeSpecName: "kube-api-access-g62zp") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "kube-api-access-g62zp". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:17:48 crc kubenswrapper[4956]: I1126 17:17:48.998195 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-config-data" (OuterVolumeSpecName: "config-data") pod "30d410b5-3f0a-4528-ba02-9c89753e575c" (UID: "30d410b5-3f0a-4528-ba02-9c89753e575c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050231 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050280 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050291 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050334 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050349 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/30d410b5-3f0a-4528-ba02-9c89753e575c-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050362 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050374 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/30d410b5-3f0a-4528-ba02-9c89753e575c-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050384 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050404 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050417 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050426 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050435 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/30d410b5-3f0a-4528-ba02-9c89753e575c-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.050445 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g62zp\" (UniqueName: \"kubernetes.io/projected/30d410b5-3f0a-4528-ba02-9c89753e575c-kube-api-access-g62zp\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.070705 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.073606 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.153920 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.153967 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.337670 4956 generic.go:334] "Generic (PLEG): container finished" podID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerID="1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc" exitCode=143 Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.337732 4956 generic.go:334] "Generic (PLEG): container finished" podID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerID="6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5" exitCode=143 Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.337781 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"30d410b5-3f0a-4528-ba02-9c89753e575c","Type":"ContainerDied","Data":"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc"} Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.337828 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"30d410b5-3f0a-4528-ba02-9c89753e575c","Type":"ContainerDied","Data":"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5"} Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.337852 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"30d410b5-3f0a-4528-ba02-9c89753e575c","Type":"ContainerDied","Data":"dad5d86daeb2a1eaa12102047fb38c27cdc13bc585247159a6fdef3c9a1207bf"} Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.337941 4956 scope.go:117] "RemoveContainer" containerID="1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.338165 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.371010 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.378310 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.379465 4956 scope.go:117] "RemoveContainer" containerID="6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.400285 4956 scope.go:117] "RemoveContainer" containerID="1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc" Nov 26 17:17:49 crc kubenswrapper[4956]: E1126 17:17:49.400793 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc\": container with ID starting with 1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc not found: ID does not exist" containerID="1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.400916 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc"} err="failed to get container status \"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc\": rpc error: code = NotFound desc = could not find container \"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc\": container with ID starting with 1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc not found: ID does not exist" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.400971 4956 scope.go:117] "RemoveContainer" containerID="6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5" Nov 26 17:17:49 crc kubenswrapper[4956]: E1126 17:17:49.401311 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5\": container with ID starting with 6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5 not found: ID does not exist" containerID="6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.401336 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5"} err="failed to get container status \"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5\": rpc error: code = NotFound desc = could not find container \"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5\": container with ID starting with 6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5 not found: ID does not exist" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.401356 4956 scope.go:117] "RemoveContainer" containerID="1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.401580 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc"} err="failed to get container status \"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc\": rpc error: code = NotFound desc = could not find container \"1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc\": container with ID starting with 1aa8d388f38340ad1f3c500115dcb24f7a8d098b602ce2abd4d73cb4610043cc not found: ID does not exist" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.401611 4956 scope.go:117] "RemoveContainer" containerID="6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.401825 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5"} err="failed to get container status \"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5\": rpc error: code = NotFound desc = could not find container \"6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5\": container with ID starting with 6cabe21bf4a8fbfe5a1f80da5a25a3fd68e91e1bb73a727310fec999c37853e5 not found: ID does not exist" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.409551 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:49 crc kubenswrapper[4956]: E1126 17:17:49.410171 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-httpd" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.410201 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-httpd" Nov 26 17:17:49 crc kubenswrapper[4956]: E1126 17:17:49.410250 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-log" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.410265 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-log" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.410480 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-httpd" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.410513 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" containerName="glance-log" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.412204 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.418798 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.425927 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.438921 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.454560 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-qt6c9" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.560325 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-run\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.560773 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-scripts\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.560802 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tz84p\" (UniqueName: \"kubernetes.io/projected/7567d08a-cded-4c9c-be35-bd7b4dac9b40-kube-api-access-tz84p\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.560834 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-httpd-run\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.560854 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561017 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-dev\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561122 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-logs\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561244 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-lib-modules\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561314 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-config-data\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561382 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561564 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561699 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561785 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-sys\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.561930 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.622393 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-dd7556475-2wzb2" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664160 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-lib-modules\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664238 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-config-data\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664277 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664331 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664322 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-lib-modules\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664364 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664408 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-sys\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664415 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664436 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664468 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-sys\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664479 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664513 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-nvme\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664556 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-run\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664599 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-scripts\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664633 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tz84p\" (UniqueName: \"kubernetes.io/projected/7567d08a-cded-4c9c-be35-bd7b4dac9b40-kube-api-access-tz84p\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664667 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-httpd-run\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664695 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664720 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-dev\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664751 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-logs\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.664889 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.665336 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-run\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.665452 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-logs\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.665731 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-httpd-run\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.665859 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.666963 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-dev\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.671970 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-config-data\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.684217 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-scripts\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.688443 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tz84p\" (UniqueName: \"kubernetes.io/projected/7567d08a-cded-4c9c-be35-bd7b4dac9b40-kube-api-access-tz84p\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.704492 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.727358 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-0\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:49 crc kubenswrapper[4956]: I1126 17:17:49.756658 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:50 crc kubenswrapper[4956]: I1126 17:17:50.195619 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:17:50 crc kubenswrapper[4956]: W1126 17:17:50.205617 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7567d08a_cded_4c9c_be35_bd7b4dac9b40.slice/crio-06543bb3dd3cac40dee87d2a5e263b00f3de947d587e0716f4435c234131eac3 WatchSource:0}: Error finding container 06543bb3dd3cac40dee87d2a5e263b00f3de947d587e0716f4435c234131eac3: Status 404 returned error can't find the container with id 06543bb3dd3cac40dee87d2a5e263b00f3de947d587e0716f4435c234131eac3 Nov 26 17:17:50 crc kubenswrapper[4956]: I1126 17:17:50.349676 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7567d08a-cded-4c9c-be35-bd7b4dac9b40","Type":"ContainerStarted","Data":"06543bb3dd3cac40dee87d2a5e263b00f3de947d587e0716f4435c234131eac3"} Nov 26 17:17:51 crc kubenswrapper[4956]: I1126 17:17:51.004572 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="30d410b5-3f0a-4528-ba02-9c89753e575c" path="/var/lib/kubelet/pods/30d410b5-3f0a-4528-ba02-9c89753e575c/volumes" Nov 26 17:17:51 crc kubenswrapper[4956]: I1126 17:17:51.362319 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7567d08a-cded-4c9c-be35-bd7b4dac9b40","Type":"ContainerStarted","Data":"0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc"} Nov 26 17:17:51 crc kubenswrapper[4956]: I1126 17:17:51.362369 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7567d08a-cded-4c9c-be35-bd7b4dac9b40","Type":"ContainerStarted","Data":"5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978"} Nov 26 17:17:51 crc kubenswrapper[4956]: I1126 17:17:51.388510 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.388481261 podStartE2EDuration="2.388481261s" podCreationTimestamp="2025-11-26 17:17:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:17:51.383085445 +0000 UTC m=+1517.079045997" watchObservedRunningTime="2025-11-26 17:17:51.388481261 +0000 UTC m=+1517.084441823" Nov 26 17:17:54 crc kubenswrapper[4956]: I1126 17:17:54.079254 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-6d556748c5-l2bl4" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.134059 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-2slwb"] Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.136316 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.148893 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2slwb"] Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.276489 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9h5wc\" (UniqueName: \"kubernetes.io/projected/47ebddca-2799-4e01-9b63-a62a2f0398fb-kube-api-access-9h5wc\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.277049 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-utilities\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.277161 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-catalog-content\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.336349 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-h5tg8"] Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.338313 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.346315 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h5tg8"] Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.379935 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-utilities\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.380016 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-catalog-content\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.380131 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9h5wc\" (UniqueName: \"kubernetes.io/projected/47ebddca-2799-4e01-9b63-a62a2f0398fb-kube-api-access-9h5wc\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.381724 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-utilities\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.381753 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-catalog-content\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.405061 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9h5wc\" (UniqueName: \"kubernetes.io/projected/47ebddca-2799-4e01-9b63-a62a2f0398fb-kube-api-access-9h5wc\") pod \"certified-operators-2slwb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.461066 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.481693 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-catalog-content\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.481979 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-utilities\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.482062 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gdtdr\" (UniqueName: \"kubernetes.io/projected/725155ee-d748-4e22-bb8e-8bbde903ecd3-kube-api-access-gdtdr\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.585090 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-catalog-content\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.585187 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-utilities\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.585256 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gdtdr\" (UniqueName: \"kubernetes.io/projected/725155ee-d748-4e22-bb8e-8bbde903ecd3-kube-api-access-gdtdr\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.586364 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-catalog-content\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.586578 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-utilities\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.607498 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gdtdr\" (UniqueName: \"kubernetes.io/projected/725155ee-d748-4e22-bb8e-8bbde903ecd3-kube-api-access-gdtdr\") pod \"community-operators-h5tg8\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.679803 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:17:56 crc kubenswrapper[4956]: I1126 17:17:56.955072 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-2slwb"] Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.154190 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-h5tg8"] Nov 26 17:17:57 crc kubenswrapper[4956]: W1126 17:17:57.158751 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod725155ee_d748_4e22_bb8e_8bbde903ecd3.slice/crio-b61eeb1ca7267d74437f6472e4e9b4334cb14b4e62a1cde86691b4af80a39d7f WatchSource:0}: Error finding container b61eeb1ca7267d74437f6472e4e9b4334cb14b4e62a1cde86691b4af80a39d7f: Status 404 returned error can't find the container with id b61eeb1ca7267d74437f6472e4e9b4334cb14b4e62a1cde86691b4af80a39d7f Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.418124 4956 generic.go:334] "Generic (PLEG): container finished" podID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerID="d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a" exitCode=0 Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.418228 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5tg8" event={"ID":"725155ee-d748-4e22-bb8e-8bbde903ecd3","Type":"ContainerDied","Data":"d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a"} Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.418348 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5tg8" event={"ID":"725155ee-d748-4e22-bb8e-8bbde903ecd3","Type":"ContainerStarted","Data":"b61eeb1ca7267d74437f6472e4e9b4334cb14b4e62a1cde86691b4af80a39d7f"} Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.420154 4956 generic.go:334] "Generic (PLEG): container finished" podID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerID="b2b23a47b9a2c4b3048cb885ea355114bf927f2af4a49ba936b57eea6ae9cce5" exitCode=0 Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.420205 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerDied","Data":"b2b23a47b9a2c4b3048cb885ea355114bf927f2af4a49ba936b57eea6ae9cce5"} Nov 26 17:17:57 crc kubenswrapper[4956]: I1126 17:17:57.420314 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerStarted","Data":"50aac5131262b31428bd79b476aab10435770d7b41eb9d943af97893cd0cee0c"} Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.330638 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-wfbpd"] Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.333558 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.341506 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfbpd"] Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.438751 4956 generic.go:334] "Generic (PLEG): container finished" podID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerID="69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e" exitCode=0 Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.439081 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5tg8" event={"ID":"725155ee-d748-4e22-bb8e-8bbde903ecd3","Type":"ContainerDied","Data":"69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e"} Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.442573 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerStarted","Data":"cc3501b013264bcd5d26ae9c5d93157b35693fcee9a6705695c27c83188f4094"} Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.536217 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-catalog-content\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.536507 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zb4ss\" (UniqueName: \"kubernetes.io/projected/0efadb03-7701-40f6-b709-67f02db4a787-kube-api-access-zb4ss\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.536561 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-utilities\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.638258 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zb4ss\" (UniqueName: \"kubernetes.io/projected/0efadb03-7701-40f6-b709-67f02db4a787-kube-api-access-zb4ss\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.638362 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-utilities\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.638436 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-catalog-content\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.638984 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-utilities\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.639053 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-catalog-content\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.660262 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zb4ss\" (UniqueName: \"kubernetes.io/projected/0efadb03-7701-40f6-b709-67f02db4a787-kube-api-access-zb4ss\") pod \"redhat-marketplace-wfbpd\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.757516 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.757591 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.816192 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.861723 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.929774 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-zzffh"] Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.931561 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.953472 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:17:59 crc kubenswrapper[4956]: I1126 17:17:59.967721 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzffh"] Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.045184 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fv587\" (UniqueName: \"kubernetes.io/projected/b4b1de34-cf17-456b-997f-ce8fa99100eb-kube-api-access-fv587\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.045297 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-catalog-content\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.045411 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-utilities\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.148356 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-catalog-content\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.149044 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-utilities\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.149075 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fv587\" (UniqueName: \"kubernetes.io/projected/b4b1de34-cf17-456b-997f-ce8fa99100eb-kube-api-access-fv587\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.149515 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-catalog-content\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.151001 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-utilities\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.180077 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fv587\" (UniqueName: \"kubernetes.io/projected/b4b1de34-cf17-456b-997f-ce8fa99100eb-kube-api-access-fv587\") pod \"redhat-operators-zzffh\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.257111 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfbpd"] Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.340065 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.492916 4956 generic.go:334] "Generic (PLEG): container finished" podID="0efadb03-7701-40f6-b709-67f02db4a787" containerID="5441e3554a1c3abe7ede68513aa05752d91de0376ccd8cdbefca62823645d4f0" exitCode=0 Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.493041 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfbpd" event={"ID":"0efadb03-7701-40f6-b709-67f02db4a787","Type":"ContainerDied","Data":"5441e3554a1c3abe7ede68513aa05752d91de0376ccd8cdbefca62823645d4f0"} Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.493610 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfbpd" event={"ID":"0efadb03-7701-40f6-b709-67f02db4a787","Type":"ContainerStarted","Data":"c6f81de748d01aab61623776a1cb196266b4659c6e8d49884a1fb5b3402a5164"} Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.522998 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5tg8" event={"ID":"725155ee-d748-4e22-bb8e-8bbde903ecd3","Type":"ContainerStarted","Data":"e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62"} Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.529498 4956 generic.go:334] "Generic (PLEG): container finished" podID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerID="cc3501b013264bcd5d26ae9c5d93157b35693fcee9a6705695c27c83188f4094" exitCode=0 Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.531479 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerDied","Data":"cc3501b013264bcd5d26ae9c5d93157b35693fcee9a6705695c27c83188f4094"} Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.531517 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.531622 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.551224 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-h5tg8" podStartSLOduration=2.076408051 podStartE2EDuration="4.551199915s" podCreationTimestamp="2025-11-26 17:17:56 +0000 UTC" firstStartedPulling="2025-11-26 17:17:57.419903079 +0000 UTC m=+1523.115863631" lastFinishedPulling="2025-11-26 17:17:59.894694943 +0000 UTC m=+1525.590655495" observedRunningTime="2025-11-26 17:18:00.545503841 +0000 UTC m=+1526.241464423" watchObservedRunningTime="2025-11-26 17:18:00.551199915 +0000 UTC m=+1526.247160467" Nov 26 17:18:00 crc kubenswrapper[4956]: W1126 17:18:00.634959 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb4b1de34_cf17_456b_997f_ce8fa99100eb.slice/crio-6b0974b1d7aba55052734d842db31878cdd55275f456c42cf54bad96232ff8d4 WatchSource:0}: Error finding container 6b0974b1d7aba55052734d842db31878cdd55275f456c42cf54bad96232ff8d4: Status 404 returned error can't find the container with id 6b0974b1d7aba55052734d842db31878cdd55275f456c42cf54bad96232ff8d4 Nov 26 17:18:00 crc kubenswrapper[4956]: I1126 17:18:00.635182 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-zzffh"] Nov 26 17:18:01 crc kubenswrapper[4956]: I1126 17:18:01.561378 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerStarted","Data":"6000a717f570741b549c70043af28fd1812f02a54b710fddf79b2c13e6f1f8e6"} Nov 26 17:18:01 crc kubenswrapper[4956]: I1126 17:18:01.566046 4956 generic.go:334] "Generic (PLEG): container finished" podID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerID="db42b20f349feaf558d74ea4e214636d5ca37c9b2ec784fc4303beae4b9f4be4" exitCode=0 Nov 26 17:18:01 crc kubenswrapper[4956]: I1126 17:18:01.567389 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzffh" event={"ID":"b4b1de34-cf17-456b-997f-ce8fa99100eb","Type":"ContainerDied","Data":"db42b20f349feaf558d74ea4e214636d5ca37c9b2ec784fc4303beae4b9f4be4"} Nov 26 17:18:01 crc kubenswrapper[4956]: I1126 17:18:01.567421 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzffh" event={"ID":"b4b1de34-cf17-456b-997f-ce8fa99100eb","Type":"ContainerStarted","Data":"6b0974b1d7aba55052734d842db31878cdd55275f456c42cf54bad96232ff8d4"} Nov 26 17:18:01 crc kubenswrapper[4956]: I1126 17:18:01.592878 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-2slwb" podStartSLOduration=1.830894186 podStartE2EDuration="5.59284131s" podCreationTimestamp="2025-11-26 17:17:56 +0000 UTC" firstStartedPulling="2025-11-26 17:17:57.424666777 +0000 UTC m=+1523.120627359" lastFinishedPulling="2025-11-26 17:18:01.186613931 +0000 UTC m=+1526.882574483" observedRunningTime="2025-11-26 17:18:01.590392219 +0000 UTC m=+1527.286352781" watchObservedRunningTime="2025-11-26 17:18:01.59284131 +0000 UTC m=+1527.288801862" Nov 26 17:18:02 crc kubenswrapper[4956]: I1126 17:18:02.577954 4956 generic.go:334] "Generic (PLEG): container finished" podID="0efadb03-7701-40f6-b709-67f02db4a787" containerID="c723914df725acdae42f3d1b06550afb389c8da0f39d708158f3d618c8aac28d" exitCode=0 Nov 26 17:18:02 crc kubenswrapper[4956]: I1126 17:18:02.578031 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfbpd" event={"ID":"0efadb03-7701-40f6-b709-67f02db4a787","Type":"ContainerDied","Data":"c723914df725acdae42f3d1b06550afb389c8da0f39d708158f3d618c8aac28d"} Nov 26 17:18:02 crc kubenswrapper[4956]: I1126 17:18:02.579081 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:18:02 crc kubenswrapper[4956]: I1126 17:18:02.579093 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:18:02 crc kubenswrapper[4956]: I1126 17:18:02.838269 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:02 crc kubenswrapper[4956]: I1126 17:18:02.840575 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:03 crc kubenswrapper[4956]: I1126 17:18:03.593724 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfbpd" event={"ID":"0efadb03-7701-40f6-b709-67f02db4a787","Type":"ContainerStarted","Data":"3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df"} Nov 26 17:18:03 crc kubenswrapper[4956]: I1126 17:18:03.597229 4956 generic.go:334] "Generic (PLEG): container finished" podID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerID="39293835d9adeb89ba8bdd70c03a5f7b4716af5e8782eb9c4fa7dc36d465ad0f" exitCode=0 Nov 26 17:18:03 crc kubenswrapper[4956]: I1126 17:18:03.597351 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzffh" event={"ID":"b4b1de34-cf17-456b-997f-ce8fa99100eb","Type":"ContainerDied","Data":"39293835d9adeb89ba8bdd70c03a5f7b4716af5e8782eb9c4fa7dc36d465ad0f"} Nov 26 17:18:03 crc kubenswrapper[4956]: I1126 17:18:03.624755 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-wfbpd" podStartSLOduration=1.889690502 podStartE2EDuration="4.624729948s" podCreationTimestamp="2025-11-26 17:17:59 +0000 UTC" firstStartedPulling="2025-11-26 17:18:00.495252142 +0000 UTC m=+1526.191212694" lastFinishedPulling="2025-11-26 17:18:03.230291588 +0000 UTC m=+1528.926252140" observedRunningTime="2025-11-26 17:18:03.62029311 +0000 UTC m=+1529.316253672" watchObservedRunningTime="2025-11-26 17:18:03.624729948 +0000 UTC m=+1529.320690500" Nov 26 17:18:04 crc kubenswrapper[4956]: I1126 17:18:04.611234 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzffh" event={"ID":"b4b1de34-cf17-456b-997f-ce8fa99100eb","Type":"ContainerStarted","Data":"a6b4c9661ffeaae79da3d2b1316bcb9642e1ed29488f45c1fa2293f3aa4922fa"} Nov 26 17:18:04 crc kubenswrapper[4956]: I1126 17:18:04.649595 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-zzffh" podStartSLOduration=2.878942917 podStartE2EDuration="5.649570967s" podCreationTimestamp="2025-11-26 17:17:59 +0000 UTC" firstStartedPulling="2025-11-26 17:18:01.568002194 +0000 UTC m=+1527.263962746" lastFinishedPulling="2025-11-26 17:18:04.338630254 +0000 UTC m=+1530.034590796" observedRunningTime="2025-11-26 17:18:04.642975927 +0000 UTC m=+1530.338936489" watchObservedRunningTime="2025-11-26 17:18:04.649570967 +0000 UTC m=+1530.345531509" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.004473 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.006167 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.007323 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.008844 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.019567 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.027755 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.156976 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157047 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-httpd-run\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157086 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-run\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157114 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-sys\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157132 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-scripts\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157155 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157176 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-nvme\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157198 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157429 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157497 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l2sj2\" (UniqueName: \"kubernetes.io/projected/26a37228-3357-4306-bfce-160f670e08ca-kube-api-access-l2sj2\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157527 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-logs\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157564 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-httpd-run\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157583 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-config-data\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157618 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157651 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-lib-modules\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157688 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-lib-modules\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157757 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157777 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-dev\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157790 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-scripts\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157845 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fjs72\" (UniqueName: \"kubernetes.io/projected/106a7b90-5a56-405e-9c99-5e0842f1a71b-kube-api-access-fjs72\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157907 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-config-data\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157926 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-dev\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157970 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.157988 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-sys\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.158071 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-logs\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.158124 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-run\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.158148 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.158174 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-nvme\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259666 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-logs\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259737 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-run\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259756 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259779 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-nvme\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259807 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259828 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-httpd-run\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259847 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-run\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259881 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-sys\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259883 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-run\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260009 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260193 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-nvme\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.259903 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-scripts\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260247 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260284 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-nvme\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260304 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260345 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260367 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l2sj2\" (UniqueName: \"kubernetes.io/projected/26a37228-3357-4306-bfce-160f670e08ca-kube-api-access-l2sj2\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260385 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-logs\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260412 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-httpd-run\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260434 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-config-data\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260460 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260485 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-lib-modules\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260507 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-lib-modules\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260528 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260548 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-dev\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260562 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-scripts\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260586 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fjs72\" (UniqueName: \"kubernetes.io/projected/106a7b90-5a56-405e-9c99-5e0842f1a71b-kube-api-access-fjs72\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260606 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-config-data\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260620 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-dev\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260642 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-sys\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260661 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.260813 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") device mount path \"/mnt/openstack/pv18\"" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.261028 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") device mount path \"/mnt/openstack/pv04\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.261365 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.261444 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-lib-modules\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.261367 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-httpd-run\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.261732 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-logs\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.261836 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-nvme\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.262107 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-lib-modules\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.262137 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") device mount path \"/mnt/openstack/pv13\"" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.262218 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-logs\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.262294 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-var-locks-brick\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.262339 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-dev\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.262803 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-sys\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.263288 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-dev\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.263330 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-run\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.263360 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-iscsi\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.263394 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-sys\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.263768 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-httpd-run\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.264025 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") device mount path \"/mnt/openstack/pv07\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.269522 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-config-data\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.273138 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-scripts\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.274192 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-config-data\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.292216 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-scripts\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.293977 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l2sj2\" (UniqueName: \"kubernetes.io/projected/26a37228-3357-4306-bfce-160f670e08ca-kube-api-access-l2sj2\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.294200 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fjs72\" (UniqueName: \"kubernetes.io/projected/106a7b90-5a56-405e-9c99-5e0842f1a71b-kube-api-access-fjs72\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.339949 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.342307 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"glance-default-single-1\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.354404 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.359426 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.366403 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"glance-default-single-2\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.641839 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:05 crc kubenswrapper[4956]: W1126 17:18:05.650291 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod26a37228_3357_4306_bfce_160f670e08ca.slice/crio-b587a7f7944d94ec7f8c7d7d5012a5cf34fa653373ec2991faf59fb0e0603e96 WatchSource:0}: Error finding container b587a7f7944d94ec7f8c7d7d5012a5cf34fa653373ec2991faf59fb0e0603e96: Status 404 returned error can't find the container with id b587a7f7944d94ec7f8c7d7d5012a5cf34fa653373ec2991faf59fb0e0603e96 Nov 26 17:18:05 crc kubenswrapper[4956]: I1126 17:18:05.663719 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.127173 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 26 17:18:06 crc kubenswrapper[4956]: W1126 17:18:06.131699 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod106a7b90_5a56_405e_9c99_5e0842f1a71b.slice/crio-28de1ddd52293a6f0fc9e425955072f42d086b67354de289e2c9384f471dbcdc WatchSource:0}: Error finding container 28de1ddd52293a6f0fc9e425955072f42d086b67354de289e2c9384f471dbcdc: Status 404 returned error can't find the container with id 28de1ddd52293a6f0fc9e425955072f42d086b67354de289e2c9384f471dbcdc Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.461495 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.462273 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.518752 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.630769 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26a37228-3357-4306-bfce-160f670e08ca","Type":"ContainerStarted","Data":"674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909"} Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.630836 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26a37228-3357-4306-bfce-160f670e08ca","Type":"ContainerStarted","Data":"cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45"} Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.630851 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26a37228-3357-4306-bfce-160f670e08ca","Type":"ContainerStarted","Data":"b587a7f7944d94ec7f8c7d7d5012a5cf34fa653373ec2991faf59fb0e0603e96"} Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.633045 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"106a7b90-5a56-405e-9c99-5e0842f1a71b","Type":"ContainerStarted","Data":"d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b"} Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.633131 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"106a7b90-5a56-405e-9c99-5e0842f1a71b","Type":"ContainerStarted","Data":"ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72"} Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.633150 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"106a7b90-5a56-405e-9c99-5e0842f1a71b","Type":"ContainerStarted","Data":"28de1ddd52293a6f0fc9e425955072f42d086b67354de289e2c9384f471dbcdc"} Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.682155 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.682233 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.715038 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=3.715011622 podStartE2EDuration="3.715011622s" podCreationTimestamp="2025-11-26 17:18:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:18:06.669621514 +0000 UTC m=+1532.365582086" watchObservedRunningTime="2025-11-26 17:18:06.715011622 +0000 UTC m=+1532.410972174" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.715315 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.736241 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-2" podStartSLOduration=3.736213223 podStartE2EDuration="3.736213223s" podCreationTimestamp="2025-11-26 17:18:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:18:06.727962405 +0000 UTC m=+1532.423922947" watchObservedRunningTime="2025-11-26 17:18:06.736213223 +0000 UTC m=+1532.432173775" Nov 26 17:18:06 crc kubenswrapper[4956]: I1126 17:18:06.780627 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:18:07 crc kubenswrapper[4956]: I1126 17:18:07.712756 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:18:09 crc kubenswrapper[4956]: I1126 17:18:09.551573 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:18:09 crc kubenswrapper[4956]: I1126 17:18:09.553082 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:18:09 crc kubenswrapper[4956]: I1126 17:18:09.956487 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:18:09 crc kubenswrapper[4956]: I1126 17:18:09.957102 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:18:10 crc kubenswrapper[4956]: I1126 17:18:10.015866 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:18:10 crc kubenswrapper[4956]: I1126 17:18:10.341250 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:10 crc kubenswrapper[4956]: I1126 17:18:10.343837 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:10 crc kubenswrapper[4956]: I1126 17:18:10.390361 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:10 crc kubenswrapper[4956]: I1126 17:18:10.717845 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:18:10 crc kubenswrapper[4956]: I1126 17:18:10.717936 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.356529 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.356952 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.386453 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.403254 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.664533 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.664649 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.698147 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.713200 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.722475 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.722999 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.723036 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:15 crc kubenswrapper[4956]: I1126 17:18:15.723065 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:17 crc kubenswrapper[4956]: I1126 17:18:17.742465 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:17 crc kubenswrapper[4956]: I1126 17:18:17.747837 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:18:17 crc kubenswrapper[4956]: I1126 17:18:17.748133 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:18:17 crc kubenswrapper[4956]: I1126 17:18:17.748234 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:18:17 crc kubenswrapper[4956]: I1126 17:18:17.915099 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:18 crc kubenswrapper[4956]: I1126 17:18:18.008601 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:18 crc kubenswrapper[4956]: I1126 17:18:18.113192 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:19 crc kubenswrapper[4956]: I1126 17:18:19.523462 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 26 17:18:19 crc kubenswrapper[4956]: I1126 17:18:19.534357 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:20 crc kubenswrapper[4956]: I1126 17:18:20.775640 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-log" containerID="cri-o://cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45" gracePeriod=30 Nov 26 17:18:20 crc kubenswrapper[4956]: I1126 17:18:20.775722 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-1" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-httpd" containerID="cri-o://674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909" gracePeriod=30 Nov 26 17:18:20 crc kubenswrapper[4956]: I1126 17:18:20.776241 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-log" containerID="cri-o://ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72" gracePeriod=30 Nov 26 17:18:20 crc kubenswrapper[4956]: I1126 17:18:20.776282 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-2" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-httpd" containerID="cri-o://d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b" gracePeriod=30 Nov 26 17:18:21 crc kubenswrapper[4956]: I1126 17:18:21.789268 4956 generic.go:334] "Generic (PLEG): container finished" podID="26a37228-3357-4306-bfce-160f670e08ca" containerID="cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45" exitCode=143 Nov 26 17:18:21 crc kubenswrapper[4956]: I1126 17:18:21.789351 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26a37228-3357-4306-bfce-160f670e08ca","Type":"ContainerDied","Data":"cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45"} Nov 26 17:18:21 crc kubenswrapper[4956]: I1126 17:18:21.792768 4956 generic.go:334] "Generic (PLEG): container finished" podID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerID="ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72" exitCode=143 Nov 26 17:18:21 crc kubenswrapper[4956]: I1126 17:18:21.792833 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"106a7b90-5a56-405e-9c99-5e0842f1a71b","Type":"ContainerDied","Data":"ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72"} Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.447791 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.505180 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540116 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-logs\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540238 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-dev\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540258 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-run\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540286 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-scripts\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540303 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-dev\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540327 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-nvme\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540359 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-httpd-run\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540384 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-iscsi\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540415 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-logs\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540447 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-httpd-run\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540481 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-config-data\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540505 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-lib-modules\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540532 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fjs72\" (UniqueName: \"kubernetes.io/projected/106a7b90-5a56-405e-9c99-5e0842f1a71b-kube-api-access-fjs72\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540560 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-scripts\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540573 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-sys\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540599 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-var-locks-brick\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540624 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540663 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-sys\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540687 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l2sj2\" (UniqueName: \"kubernetes.io/projected/26a37228-3357-4306-bfce-160f670e08ca-kube-api-access-l2sj2\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540710 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-config-data\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540732 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-lib-modules\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540748 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-run\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540766 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-nvme\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540785 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-iscsi\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540816 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540835 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") pod \"106a7b90-5a56-405e-9c99-5e0842f1a71b\" (UID: \"106a7b90-5a56-405e-9c99-5e0842f1a71b\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540855 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-var-locks-brick\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.540902 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"26a37228-3357-4306-bfce-160f670e08ca\" (UID: \"26a37228-3357-4306-bfce-160f670e08ca\") " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.541267 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.541626 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-logs" (OuterVolumeSpecName: "logs") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.541662 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-dev" (OuterVolumeSpecName: "dev") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.541690 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-run" (OuterVolumeSpecName: "run") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542344 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542427 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-dev" (OuterVolumeSpecName: "dev") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542458 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542604 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542705 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542793 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-logs" (OuterVolumeSpecName: "logs") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542827 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542910 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.542997 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.543051 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-sys" (OuterVolumeSpecName: "sys") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.543103 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.543118 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.543828 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-run" (OuterVolumeSpecName: "run") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.543970 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-sys" (OuterVolumeSpecName: "sys") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.550359 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage07-crc" (OuterVolumeSpecName: "glance") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "local-storage07-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.550574 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance-cache") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.551673 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage13-crc" (OuterVolumeSpecName: "glance-cache") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "local-storage13-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.552221 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/106a7b90-5a56-405e-9c99-5e0842f1a71b-kube-api-access-fjs72" (OuterVolumeSpecName: "kube-api-access-fjs72") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "kube-api-access-fjs72". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.552380 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/26a37228-3357-4306-bfce-160f670e08ca-kube-api-access-l2sj2" (OuterVolumeSpecName: "kube-api-access-l2sj2") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "kube-api-access-l2sj2". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.552570 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-scripts" (OuterVolumeSpecName: "scripts") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.555007 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage18-crc" (OuterVolumeSpecName: "glance") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "local-storage18-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.556616 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-scripts" (OuterVolumeSpecName: "scripts") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.592691 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-config-data" (OuterVolumeSpecName: "config-data") pod "26a37228-3357-4306-bfce-160f670e08ca" (UID: "26a37228-3357-4306-bfce-160f670e08ca"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.595525 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-config-data" (OuterVolumeSpecName: "config-data") pod "106a7b90-5a56-405e-9c99-5e0842f1a71b" (UID: "106a7b90-5a56-405e-9c99-5e0842f1a71b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643306 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643355 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643405 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643423 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643435 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643454 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643467 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643501 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643511 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643521 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643532 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643542 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643551 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/106a7b90-5a56-405e-9c99-5e0842f1a71b-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643796 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643820 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643832 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/26a37228-3357-4306-bfce-160f670e08ca-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643841 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643849 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643892 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fjs72\" (UniqueName: \"kubernetes.io/projected/106a7b90-5a56-405e-9c99-5e0842f1a71b-kube-api-access-fjs72\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643903 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/106a7b90-5a56-405e-9c99-5e0842f1a71b-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643912 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643920 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643976 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643987 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.643997 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l2sj2\" (UniqueName: \"kubernetes.io/projected/26a37228-3357-4306-bfce-160f670e08ca-kube-api-access-l2sj2\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.644006 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/26a37228-3357-4306-bfce-160f670e08ca-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.644015 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/106a7b90-5a56-405e-9c99-5e0842f1a71b-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.644024 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/26a37228-3357-4306-bfce-160f670e08ca-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.657901 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage13-crc" (UniqueName: "kubernetes.io/local-volume/local-storage13-crc") on node "crc" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.658557 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.667670 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage07-crc" (UniqueName: "kubernetes.io/local-volume/local-storage07-crc") on node "crc" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.673907 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage18-crc" (UniqueName: "kubernetes.io/local-volume/local-storage18-crc") on node "crc" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.746224 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage13-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage13-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.746280 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage18-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage18-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.746294 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.746308 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.829088 4956 generic.go:334] "Generic (PLEG): container finished" podID="26a37228-3357-4306-bfce-160f670e08ca" containerID="674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909" exitCode=0 Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.829216 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.829209 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26a37228-3357-4306-bfce-160f670e08ca","Type":"ContainerDied","Data":"674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909"} Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.829362 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"26a37228-3357-4306-bfce-160f670e08ca","Type":"ContainerDied","Data":"b587a7f7944d94ec7f8c7d7d5012a5cf34fa653373ec2991faf59fb0e0603e96"} Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.829385 4956 scope.go:117] "RemoveContainer" containerID="674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.832140 4956 generic.go:334] "Generic (PLEG): container finished" podID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerID="d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b" exitCode=0 Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.832172 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"106a7b90-5a56-405e-9c99-5e0842f1a71b","Type":"ContainerDied","Data":"d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b"} Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.832191 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-2" event={"ID":"106a7b90-5a56-405e-9c99-5e0842f1a71b","Type":"ContainerDied","Data":"28de1ddd52293a6f0fc9e425955072f42d086b67354de289e2c9384f471dbcdc"} Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.832263 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-2" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.865607 4956 scope.go:117] "RemoveContainer" containerID="cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.894517 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.909957 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-2"] Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.910918 4956 scope.go:117] "RemoveContainer" containerID="674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909" Nov 26 17:18:24 crc kubenswrapper[4956]: E1126 17:18:24.911718 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909\": container with ID starting with 674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909 not found: ID does not exist" containerID="674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.911782 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909"} err="failed to get container status \"674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909\": rpc error: code = NotFound desc = could not find container \"674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909\": container with ID starting with 674c21ef4663b8087e1e81ab2d19fa7f735bcfb3156eba043b76581520f49909 not found: ID does not exist" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.911836 4956 scope.go:117] "RemoveContainer" containerID="cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45" Nov 26 17:18:24 crc kubenswrapper[4956]: E1126 17:18:24.913032 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45\": container with ID starting with cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45 not found: ID does not exist" containerID="cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.913063 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45"} err="failed to get container status \"cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45\": rpc error: code = NotFound desc = could not find container \"cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45\": container with ID starting with cf9fa9ab24f7780188b16d9c771d8534a100cf605a31ea224f69d015312eec45 not found: ID does not exist" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.913077 4956 scope.go:117] "RemoveContainer" containerID="d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.919976 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.929215 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.945150 4956 scope.go:117] "RemoveContainer" containerID="ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.980069 4956 scope.go:117] "RemoveContainer" containerID="d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b" Nov 26 17:18:24 crc kubenswrapper[4956]: E1126 17:18:24.981335 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b\": container with ID starting with d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b not found: ID does not exist" containerID="d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.981381 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b"} err="failed to get container status \"d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b\": rpc error: code = NotFound desc = could not find container \"d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b\": container with ID starting with d367a783aa192f1576bb0a1466835390bc1d7dc353f3f77c1dfe42c0bd25cf2b not found: ID does not exist" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.981409 4956 scope.go:117] "RemoveContainer" containerID="ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72" Nov 26 17:18:24 crc kubenswrapper[4956]: E1126 17:18:24.981820 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72\": container with ID starting with ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72 not found: ID does not exist" containerID="ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72" Nov 26 17:18:24 crc kubenswrapper[4956]: I1126 17:18:24.981940 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72"} err="failed to get container status \"ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72\": rpc error: code = NotFound desc = could not find container \"ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72\": container with ID starting with ff851a1b7a4e23fb7c35b5b142e425dbd27203905184dd671f4dcbcabc28da72 not found: ID does not exist" Nov 26 17:18:25 crc kubenswrapper[4956]: I1126 17:18:25.013813 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" path="/var/lib/kubelet/pods/106a7b90-5a56-405e-9c99-5e0842f1a71b/volumes" Nov 26 17:18:25 crc kubenswrapper[4956]: I1126 17:18:25.014751 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="26a37228-3357-4306-bfce-160f670e08ca" path="/var/lib/kubelet/pods/26a37228-3357-4306-bfce-160f670e08ca/volumes" Nov 26 17:18:25 crc kubenswrapper[4956]: I1126 17:18:25.947310 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:18:25 crc kubenswrapper[4956]: I1126 17:18:25.947605 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-log" containerID="cri-o://5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978" gracePeriod=30 Nov 26 17:18:25 crc kubenswrapper[4956]: I1126 17:18:25.948145 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-httpd" containerID="cri-o://0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc" gracePeriod=30 Nov 26 17:18:26 crc kubenswrapper[4956]: I1126 17:18:26.862091 4956 generic.go:334] "Generic (PLEG): container finished" podID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerID="5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978" exitCode=143 Nov 26 17:18:26 crc kubenswrapper[4956]: I1126 17:18:26.862179 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7567d08a-cded-4c9c-be35-bd7b4dac9b40","Type":"ContainerDied","Data":"5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978"} Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.510383 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640043 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tz84p\" (UniqueName: \"kubernetes.io/projected/7567d08a-cded-4c9c-be35-bd7b4dac9b40-kube-api-access-tz84p\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640167 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-config-data\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640193 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-var-locks-brick\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640228 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-logs\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640362 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-iscsi\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640338 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640393 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641387 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-run\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640429 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641422 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-scripts\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.640733 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-logs" (OuterVolumeSpecName: "logs") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641461 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-lib-modules\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641481 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-run" (OuterVolumeSpecName: "run") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641508 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-nvme\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641536 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641572 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-httpd-run\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641603 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-sys\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.641626 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-dev\") pod \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\" (UID: \"7567d08a-cded-4c9c-be35-bd7b4dac9b40\") " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642139 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642175 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642193 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642209 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642244 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642286 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-sys" (OuterVolumeSpecName: "sys") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642260 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-dev" (OuterVolumeSpecName: "dev") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642332 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.642617 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.647805 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "glance-cache") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.650132 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-scripts" (OuterVolumeSpecName: "scripts") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.650314 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7567d08a-cded-4c9c-be35-bd7b4dac9b40-kube-api-access-tz84p" (OuterVolumeSpecName: "kube-api-access-tz84p") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "kube-api-access-tz84p". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.653162 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage19-crc" (OuterVolumeSpecName: "glance") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "local-storage19-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.697589 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-config-data" (OuterVolumeSpecName: "config-data") pod "7567d08a-cded-4c9c-be35-bd7b4dac9b40" (UID: "7567d08a-cded-4c9c-be35-bd7b4dac9b40"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746181 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746228 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746278 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746290 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/7567d08a-cded-4c9c-be35-bd7b4dac9b40-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746302 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746311 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/7567d08a-cded-4c9c-be35-bd7b4dac9b40-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746322 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tz84p\" (UniqueName: \"kubernetes.io/projected/7567d08a-cded-4c9c-be35-bd7b4dac9b40-kube-api-access-tz84p\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746333 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746348 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" " Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.746360 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/7567d08a-cded-4c9c-be35-bd7b4dac9b40-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.791304 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.795038 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage19-crc" (UniqueName: "kubernetes.io/local-volume/local-storage19-crc") on node "crc" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.847515 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.847566 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.895255 4956 generic.go:334] "Generic (PLEG): container finished" podID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerID="0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc" exitCode=0 Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.895322 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7567d08a-cded-4c9c-be35-bd7b4dac9b40","Type":"ContainerDied","Data":"0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc"} Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.895334 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.895357 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"7567d08a-cded-4c9c-be35-bd7b4dac9b40","Type":"ContainerDied","Data":"06543bb3dd3cac40dee87d2a5e263b00f3de947d587e0716f4435c234131eac3"} Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.895378 4956 scope.go:117] "RemoveContainer" containerID="0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.931520 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.941697 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.942346 4956 scope.go:117] "RemoveContainer" containerID="5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.966448 4956 scope.go:117] "RemoveContainer" containerID="0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc" Nov 26 17:18:29 crc kubenswrapper[4956]: E1126 17:18:29.967435 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc\": container with ID starting with 0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc not found: ID does not exist" containerID="0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.967474 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc"} err="failed to get container status \"0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc\": rpc error: code = NotFound desc = could not find container \"0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc\": container with ID starting with 0738a981af35345e2bfe88d806aa54ea6fc974ac37709331408953829c4bf6fc not found: ID does not exist" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.967506 4956 scope.go:117] "RemoveContainer" containerID="5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978" Nov 26 17:18:29 crc kubenswrapper[4956]: E1126 17:18:29.968130 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978\": container with ID starting with 5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978 not found: ID does not exist" containerID="5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978" Nov 26 17:18:29 crc kubenswrapper[4956]: I1126 17:18:29.968196 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978"} err="failed to get container status \"5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978\": rpc error: code = NotFound desc = could not find container \"5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978\": container with ID starting with 5056886716b436e1fa057a7fa864f5af6d5588929f1f08f917e7b39347757978 not found: ID does not exist" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.008456 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" path="/var/lib/kubelet/pods/7567d08a-cded-4c9c-be35-bd7b4dac9b40/volumes" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.258013 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hc2md"] Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.265556 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-sync-hc2md"] Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293356 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glancef567-account-delete-ktvvs"] Nov 26 17:18:31 crc kubenswrapper[4956]: E1126 17:18:31.293769 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293790 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: E1126 17:18:31.293810 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293819 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: E1126 17:18:31.293833 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293840 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: E1126 17:18:31.293873 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293882 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: E1126 17:18:31.293909 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293918 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: E1126 17:18:31.293936 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.293944 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294102 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294114 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294120 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="106a7b90-5a56-405e-9c99-5e0842f1a71b" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294132 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="7567d08a-cded-4c9c-be35-bd7b4dac9b40" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294149 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-log" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294165 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="26a37228-3357-4306-bfce-160f670e08ca" containerName="glance-httpd" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.294729 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.306937 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef567-account-delete-ktvvs"] Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.479532 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cdjtx\" (UniqueName: \"kubernetes.io/projected/de419132-d302-4aa2-bce5-22704eaec943-kube-api-access-cdjtx\") pod \"glancef567-account-delete-ktvvs\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.480027 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de419132-d302-4aa2-bce5-22704eaec943-operator-scripts\") pod \"glancef567-account-delete-ktvvs\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.581906 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cdjtx\" (UniqueName: \"kubernetes.io/projected/de419132-d302-4aa2-bce5-22704eaec943-kube-api-access-cdjtx\") pod \"glancef567-account-delete-ktvvs\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.582080 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de419132-d302-4aa2-bce5-22704eaec943-operator-scripts\") pod \"glancef567-account-delete-ktvvs\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.583141 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de419132-d302-4aa2-bce5-22704eaec943-operator-scripts\") pod \"glancef567-account-delete-ktvvs\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.605740 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cdjtx\" (UniqueName: \"kubernetes.io/projected/de419132-d302-4aa2-bce5-22704eaec943-kube-api-access-cdjtx\") pod \"glancef567-account-delete-ktvvs\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.611598 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.871098 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glancef567-account-delete-ktvvs"] Nov 26 17:18:31 crc kubenswrapper[4956]: I1126 17:18:31.920152 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" event={"ID":"de419132-d302-4aa2-bce5-22704eaec943","Type":"ContainerStarted","Data":"a3b827df25d2d4dce02c87dc3f0f870e6526b8fa71cdf04940968b5f19d256f5"} Nov 26 17:18:32 crc kubenswrapper[4956]: I1126 17:18:32.929300 4956 generic.go:334] "Generic (PLEG): container finished" podID="de419132-d302-4aa2-bce5-22704eaec943" containerID="1266462912a885656e1f047d08ab65c6e573ae1f47cc909452e5a33df74a577c" exitCode=0 Nov 26 17:18:32 crc kubenswrapper[4956]: I1126 17:18:32.929411 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" event={"ID":"de419132-d302-4aa2-bce5-22704eaec943","Type":"ContainerDied","Data":"1266462912a885656e1f047d08ab65c6e573ae1f47cc909452e5a33df74a577c"} Nov 26 17:18:33 crc kubenswrapper[4956]: I1126 17:18:33.005943 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fdba013-2bd9-4c44-be91-a2490b083e1c" path="/var/lib/kubelet/pods/2fdba013-2bd9-4c44-be91-a2490b083e1c/volumes" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.337901 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.434489 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cdjtx\" (UniqueName: \"kubernetes.io/projected/de419132-d302-4aa2-bce5-22704eaec943-kube-api-access-cdjtx\") pod \"de419132-d302-4aa2-bce5-22704eaec943\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.434612 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de419132-d302-4aa2-bce5-22704eaec943-operator-scripts\") pod \"de419132-d302-4aa2-bce5-22704eaec943\" (UID: \"de419132-d302-4aa2-bce5-22704eaec943\") " Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.437012 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/de419132-d302-4aa2-bce5-22704eaec943-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "de419132-d302-4aa2-bce5-22704eaec943" (UID: "de419132-d302-4aa2-bce5-22704eaec943"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.443021 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de419132-d302-4aa2-bce5-22704eaec943-kube-api-access-cdjtx" (OuterVolumeSpecName: "kube-api-access-cdjtx") pod "de419132-d302-4aa2-bce5-22704eaec943" (UID: "de419132-d302-4aa2-bce5-22704eaec943"). InnerVolumeSpecName "kube-api-access-cdjtx". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.536818 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/de419132-d302-4aa2-bce5-22704eaec943-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.536872 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cdjtx\" (UniqueName: \"kubernetes.io/projected/de419132-d302-4aa2-bce5-22704eaec943-kube-api-access-cdjtx\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.971385 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.971289 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glancef567-account-delete-ktvvs" event={"ID":"de419132-d302-4aa2-bce5-22704eaec943","Type":"ContainerDied","Data":"a3b827df25d2d4dce02c87dc3f0f870e6526b8fa71cdf04940968b5f19d256f5"} Nov 26 17:18:34 crc kubenswrapper[4956]: I1126 17:18:34.972108 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a3b827df25d2d4dce02c87dc3f0f870e6526b8fa71cdf04940968b5f19d256f5" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.550559 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:18:35 crc kubenswrapper[4956]: E1126 17:18:35.551431 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de419132-d302-4aa2-bce5-22704eaec943" containerName="mariadb-account-delete" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.551451 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="de419132-d302-4aa2-bce5-22704eaec943" containerName="mariadb-account-delete" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.551612 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="de419132-d302-4aa2-bce5-22704eaec943" containerName="mariadb-account-delete" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.552267 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.555037 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-config" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.555100 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"openstack-config-secret" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.555808 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"glance-kuttl-tests"/"openstack-scripts-9db6gc427h" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.561221 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"default-dockercfg-7fph6" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.575477 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.663717 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-config\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.663794 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-scripts\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.664668 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-config-secret\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.664775 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqpfw\" (UniqueName: \"kubernetes.io/projected/c193b63d-c72a-4f03-a669-2fd125152da6-kube-api-access-vqpfw\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.766394 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-config-secret\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.766459 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqpfw\" (UniqueName: \"kubernetes.io/projected/c193b63d-c72a-4f03-a669-2fd125152da6-kube-api-access-vqpfw\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.766516 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-config\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.766542 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-scripts\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.768615 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-config\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.768620 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-scripts\" (UniqueName: \"kubernetes.io/configmap/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-scripts\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.772426 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/c193b63d-c72a-4f03-a669-2fd125152da6-openstack-config-secret\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.784569 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqpfw\" (UniqueName: \"kubernetes.io/projected/c193b63d-c72a-4f03-a669-2fd125152da6-kube-api-access-vqpfw\") pod \"openstackclient\" (UID: \"c193b63d-c72a-4f03-a669-2fd125152da6\") " pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:35 crc kubenswrapper[4956]: I1126 17:18:35.877334 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/openstackclient" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.322279 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-db-create-fmlrm"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.331068 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-db-create-fmlrm"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.338144 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-f567-account-create-update-j8wzt"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.347137 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glancef567-account-delete-ktvvs"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.356123 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-f567-account-create-update-j8wzt"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.363038 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glancef567-account-delete-ktvvs"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.390040 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/openstackclient"] Nov 26 17:18:36 crc kubenswrapper[4956]: W1126 17:18:36.394316 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc193b63d_c72a_4f03_a669_2fd125152da6.slice/crio-c5833ae265e05f27fd70dd183af3bab92ec686f601b31c754b6a5c7c626ba140 WatchSource:0}: Error finding container c5833ae265e05f27fd70dd183af3bab92ec686f601b31c754b6a5c7c626ba140: Status 404 returned error can't find the container with id c5833ae265e05f27fd70dd183af3bab92ec686f601b31c754b6a5c7c626ba140 Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.567628 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.569640 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.574408 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-db-secret" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.592420 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-create-cz42g"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.594033 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.600467 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.606903 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-cz42g"] Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.681656 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-operator-scripts\") pod \"glance-9e5d-account-create-update-qhfkw\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.681740 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d87nt\" (UniqueName: \"kubernetes.io/projected/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-kube-api-access-d87nt\") pod \"glance-9e5d-account-create-update-qhfkw\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.681826 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-operator-scripts\") pod \"glance-db-create-cz42g\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.681905 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f9tkn\" (UniqueName: \"kubernetes.io/projected/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-kube-api-access-f9tkn\") pod \"glance-db-create-cz42g\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.783351 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d87nt\" (UniqueName: \"kubernetes.io/projected/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-kube-api-access-d87nt\") pod \"glance-9e5d-account-create-update-qhfkw\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.783456 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-operator-scripts\") pod \"glance-db-create-cz42g\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.783521 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f9tkn\" (UniqueName: \"kubernetes.io/projected/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-kube-api-access-f9tkn\") pod \"glance-db-create-cz42g\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.783570 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-operator-scripts\") pod \"glance-9e5d-account-create-update-qhfkw\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.784664 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-operator-scripts\") pod \"glance-9e5d-account-create-update-qhfkw\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.784884 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-operator-scripts\") pod \"glance-db-create-cz42g\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.802944 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f9tkn\" (UniqueName: \"kubernetes.io/projected/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-kube-api-access-f9tkn\") pod \"glance-db-create-cz42g\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.805911 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d87nt\" (UniqueName: \"kubernetes.io/projected/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-kube-api-access-d87nt\") pod \"glance-9e5d-account-create-update-qhfkw\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.909044 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:36 crc kubenswrapper[4956]: I1126 17:18:36.920512 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.018217 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2c82e33c-c999-4ac1-a9a9-770948eab705" path="/var/lib/kubelet/pods/2c82e33c-c999-4ac1-a9a9-770948eab705/volumes" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.019208 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d275f55f-c0e8-47e2-bf81-1621f1d66a76" path="/var/lib/kubelet/pods/d275f55f-c0e8-47e2-bf81-1621f1d66a76/volumes" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.021357 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de419132-d302-4aa2-bce5-22704eaec943" path="/var/lib/kubelet/pods/de419132-d302-4aa2-bce5-22704eaec943/volumes" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.023255 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"c193b63d-c72a-4f03-a669-2fd125152da6","Type":"ContainerStarted","Data":"c7dbaf3369d0e72a7ccd316b75c3a937e613b02c648bd8f95263d406164ac2b7"} Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.023316 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/openstackclient" event={"ID":"c193b63d-c72a-4f03-a669-2fd125152da6","Type":"ContainerStarted","Data":"c5833ae265e05f27fd70dd183af3bab92ec686f601b31c754b6a5c7c626ba140"} Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.041527 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/openstackclient" podStartSLOduration=2.041504683 podStartE2EDuration="2.041504683s" podCreationTimestamp="2025-11-26 17:18:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:18:37.037812326 +0000 UTC m=+1562.733772878" watchObservedRunningTime="2025-11-26 17:18:37.041504683 +0000 UTC m=+1562.737465235" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.126689 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h5tg8"] Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.127687 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-h5tg8" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="registry-server" containerID="cri-o://e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62" gracePeriod=2 Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.242693 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-create-cz42g"] Nov 26 17:18:37 crc kubenswrapper[4956]: W1126 17:18:37.250319 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6ee6b955_28a3_45e4_ba60_154ddd97ed1d.slice/crio-eb7694fbb192f36319a7279d647b14f1dfd270b23609b544bd3d422f60ae64a3 WatchSource:0}: Error finding container eb7694fbb192f36319a7279d647b14f1dfd270b23609b544bd3d422f60ae64a3: Status 404 returned error can't find the container with id eb7694fbb192f36319a7279d647b14f1dfd270b23609b544bd3d422f60ae64a3 Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.391627 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw"] Nov 26 17:18:37 crc kubenswrapper[4956]: W1126 17:18:37.430736 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poded6c8be2_8540_4937_8ee8_7209c72b1fbe.slice/crio-a65223e7ae6d6aebf7d7d488e42155fb98396201d433dfab885bcc6cba7f8671 WatchSource:0}: Error finding container a65223e7ae6d6aebf7d7d488e42155fb98396201d433dfab885bcc6cba7f8671: Status 404 returned error can't find the container with id a65223e7ae6d6aebf7d7d488e42155fb98396201d433dfab885bcc6cba7f8671 Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.611417 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.699573 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gdtdr\" (UniqueName: \"kubernetes.io/projected/725155ee-d748-4e22-bb8e-8bbde903ecd3-kube-api-access-gdtdr\") pod \"725155ee-d748-4e22-bb8e-8bbde903ecd3\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.700280 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-utilities\") pod \"725155ee-d748-4e22-bb8e-8bbde903ecd3\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.700450 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-catalog-content\") pod \"725155ee-d748-4e22-bb8e-8bbde903ecd3\" (UID: \"725155ee-d748-4e22-bb8e-8bbde903ecd3\") " Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.702684 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-utilities" (OuterVolumeSpecName: "utilities") pod "725155ee-d748-4e22-bb8e-8bbde903ecd3" (UID: "725155ee-d748-4e22-bb8e-8bbde903ecd3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.722763 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/725155ee-d748-4e22-bb8e-8bbde903ecd3-kube-api-access-gdtdr" (OuterVolumeSpecName: "kube-api-access-gdtdr") pod "725155ee-d748-4e22-bb8e-8bbde903ecd3" (UID: "725155ee-d748-4e22-bb8e-8bbde903ecd3"). InnerVolumeSpecName "kube-api-access-gdtdr". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.763856 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "725155ee-d748-4e22-bb8e-8bbde903ecd3" (UID: "725155ee-d748-4e22-bb8e-8bbde903ecd3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.802134 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.802182 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gdtdr\" (UniqueName: \"kubernetes.io/projected/725155ee-d748-4e22-bb8e-8bbde903ecd3-kube-api-access-gdtdr\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:37 crc kubenswrapper[4956]: I1126 17:18:37.802200 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/725155ee-d748-4e22-bb8e-8bbde903ecd3-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.033730 4956 generic.go:334] "Generic (PLEG): container finished" podID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerID="e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62" exitCode=0 Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.033838 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-h5tg8" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.033854 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5tg8" event={"ID":"725155ee-d748-4e22-bb8e-8bbde903ecd3","Type":"ContainerDied","Data":"e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62"} Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.033968 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-h5tg8" event={"ID":"725155ee-d748-4e22-bb8e-8bbde903ecd3","Type":"ContainerDied","Data":"b61eeb1ca7267d74437f6472e4e9b4334cb14b4e62a1cde86691b4af80a39d7f"} Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.033991 4956 scope.go:117] "RemoveContainer" containerID="e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.035734 4956 generic.go:334] "Generic (PLEG): container finished" podID="ed6c8be2-8540-4937-8ee8-7209c72b1fbe" containerID="2c888b4cd79c1bd6af8f62c3afef23ec80c77d40d8a6bc5bdcb0b7f2f67b8bb6" exitCode=0 Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.035791 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" event={"ID":"ed6c8be2-8540-4937-8ee8-7209c72b1fbe","Type":"ContainerDied","Data":"2c888b4cd79c1bd6af8f62c3afef23ec80c77d40d8a6bc5bdcb0b7f2f67b8bb6"} Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.035822 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" event={"ID":"ed6c8be2-8540-4937-8ee8-7209c72b1fbe","Type":"ContainerStarted","Data":"a65223e7ae6d6aebf7d7d488e42155fb98396201d433dfab885bcc6cba7f8671"} Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.047930 4956 generic.go:334] "Generic (PLEG): container finished" podID="6ee6b955-28a3-45e4-ba60-154ddd97ed1d" containerID="e9e20be72451cdfdec90ea16676a375bca7031115e4b04a76a141f2cb4183ac3" exitCode=0 Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.048118 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cz42g" event={"ID":"6ee6b955-28a3-45e4-ba60-154ddd97ed1d","Type":"ContainerDied","Data":"e9e20be72451cdfdec90ea16676a375bca7031115e4b04a76a141f2cb4183ac3"} Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.048170 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cz42g" event={"ID":"6ee6b955-28a3-45e4-ba60-154ddd97ed1d","Type":"ContainerStarted","Data":"eb7694fbb192f36319a7279d647b14f1dfd270b23609b544bd3d422f60ae64a3"} Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.059825 4956 scope.go:117] "RemoveContainer" containerID="69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.095433 4956 scope.go:117] "RemoveContainer" containerID="d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.124263 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-h5tg8"] Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.144724 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-h5tg8"] Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.149987 4956 scope.go:117] "RemoveContainer" containerID="e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62" Nov 26 17:18:38 crc kubenswrapper[4956]: E1126 17:18:38.151948 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62\": container with ID starting with e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62 not found: ID does not exist" containerID="e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.152004 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62"} err="failed to get container status \"e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62\": rpc error: code = NotFound desc = could not find container \"e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62\": container with ID starting with e48a057881969cc2bdd5db547609ca2a7ab01079e5f67906ea9c068f942dcf62 not found: ID does not exist" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.152039 4956 scope.go:117] "RemoveContainer" containerID="69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e" Nov 26 17:18:38 crc kubenswrapper[4956]: E1126 17:18:38.152734 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e\": container with ID starting with 69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e not found: ID does not exist" containerID="69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.152793 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e"} err="failed to get container status \"69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e\": rpc error: code = NotFound desc = could not find container \"69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e\": container with ID starting with 69566491581ca694855e9789f19a46e4b7ce7a92dc1c25a8a0c7f99c72640e9e not found: ID does not exist" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.152832 4956 scope.go:117] "RemoveContainer" containerID="d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a" Nov 26 17:18:38 crc kubenswrapper[4956]: E1126 17:18:38.153184 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a\": container with ID starting with d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a not found: ID does not exist" containerID="d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a" Nov 26 17:18:38 crc kubenswrapper[4956]: I1126 17:18:38.153213 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a"} err="failed to get container status \"d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a\": rpc error: code = NotFound desc = could not find container \"d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a\": container with ID starting with d9f79a9bcee92d123abaeed9172a055314468991ed3c4abfbdb2c0c8a2b1818a not found: ID does not exist" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.013187 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" path="/var/lib/kubelet/pods/725155ee-d748-4e22-bb8e-8bbde903ecd3/volumes" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.547859 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.550874 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.550930 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.550982 4956 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.551589 4956 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186"} pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.551649 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" containerID="cri-o://3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" gracePeriod=600 Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.554107 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.637625 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f9tkn\" (UniqueName: \"kubernetes.io/projected/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-kube-api-access-f9tkn\") pod \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.637775 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d87nt\" (UniqueName: \"kubernetes.io/projected/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-kube-api-access-d87nt\") pod \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.637835 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-operator-scripts\") pod \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\" (UID: \"6ee6b955-28a3-45e4-ba60-154ddd97ed1d\") " Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.638084 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-operator-scripts\") pod \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\" (UID: \"ed6c8be2-8540-4937-8ee8-7209c72b1fbe\") " Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.638664 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "ed6c8be2-8540-4937-8ee8-7209c72b1fbe" (UID: "ed6c8be2-8540-4937-8ee8-7209c72b1fbe"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.638805 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-operator-scripts" (OuterVolumeSpecName: "operator-scripts") pod "6ee6b955-28a3-45e4-ba60-154ddd97ed1d" (UID: "6ee6b955-28a3-45e4-ba60-154ddd97ed1d"). InnerVolumeSpecName "operator-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.644398 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-kube-api-access-f9tkn" (OuterVolumeSpecName: "kube-api-access-f9tkn") pod "6ee6b955-28a3-45e4-ba60-154ddd97ed1d" (UID: "6ee6b955-28a3-45e4-ba60-154ddd97ed1d"). InnerVolumeSpecName "kube-api-access-f9tkn". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.644586 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-kube-api-access-d87nt" (OuterVolumeSpecName: "kube-api-access-d87nt") pod "ed6c8be2-8540-4937-8ee8-7209c72b1fbe" (UID: "ed6c8be2-8540-4937-8ee8-7209c72b1fbe"). InnerVolumeSpecName "kube-api-access-d87nt". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:39 crc kubenswrapper[4956]: E1126 17:18:39.689591 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.725114 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfbpd"] Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.725423 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-wfbpd" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="registry-server" containerID="cri-o://3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df" gracePeriod=2 Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.739774 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d87nt\" (UniqueName: \"kubernetes.io/projected/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-kube-api-access-d87nt\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.739827 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.739843 4956 reconciler_common.go:293] "Volume detached for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ed6c8be2-8540-4937-8ee8-7209c72b1fbe-operator-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:39 crc kubenswrapper[4956]: I1126 17:18:39.739875 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f9tkn\" (UniqueName: \"kubernetes.io/projected/6ee6b955-28a3-45e4-ba60-154ddd97ed1d-kube-api-access-f9tkn\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:39 crc kubenswrapper[4956]: E1126 17:18:39.805016 4956 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0efadb03_7701_40f6_b709_67f02db4a787.slice/crio-3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df.scope\": RecentStats: unable to find data in memory cache]" Nov 26 17:18:39 crc kubenswrapper[4956]: E1126 17:18:39.959063 4956 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df is running failed: container process not found" containerID="3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df" cmd=["grpc_health_probe","-addr=:50051"] Nov 26 17:18:39 crc kubenswrapper[4956]: E1126 17:18:39.960255 4956 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df is running failed: container process not found" containerID="3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df" cmd=["grpc_health_probe","-addr=:50051"] Nov 26 17:18:39 crc kubenswrapper[4956]: E1126 17:18:39.961373 4956 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df is running failed: container process not found" containerID="3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df" cmd=["grpc_health_probe","-addr=:50051"] Nov 26 17:18:39 crc kubenswrapper[4956]: E1126 17:18:39.961435 4956 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of 3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-wfbpd" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="registry-server" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.071252 4956 generic.go:334] "Generic (PLEG): container finished" podID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" exitCode=0 Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.071331 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerDied","Data":"3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186"} Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.071381 4956 scope.go:117] "RemoveContainer" containerID="6afa66b60ccff62ba0cee9768e910db0d5a65279fb098588258f5b5dc187f77e" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.072152 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:18:40 crc kubenswrapper[4956]: E1126 17:18:40.072434 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.078183 4956 generic.go:334] "Generic (PLEG): container finished" podID="0efadb03-7701-40f6-b709-67f02db4a787" containerID="3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df" exitCode=0 Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.078245 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfbpd" event={"ID":"0efadb03-7701-40f6-b709-67f02db4a787","Type":"ContainerDied","Data":"3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df"} Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.080336 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" event={"ID":"ed6c8be2-8540-4937-8ee8-7209c72b1fbe","Type":"ContainerDied","Data":"a65223e7ae6d6aebf7d7d488e42155fb98396201d433dfab885bcc6cba7f8671"} Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.080383 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a65223e7ae6d6aebf7d7d488e42155fb98396201d433dfab885bcc6cba7f8671" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.080384 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-9e5d-account-create-update-qhfkw" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.082342 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-create-cz42g" event={"ID":"6ee6b955-28a3-45e4-ba60-154ddd97ed1d","Type":"ContainerDied","Data":"eb7694fbb192f36319a7279d647b14f1dfd270b23609b544bd3d422f60ae64a3"} Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.082369 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eb7694fbb192f36319a7279d647b14f1dfd270b23609b544bd3d422f60ae64a3" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.082396 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-create-cz42g" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.150152 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.249856 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-utilities\") pod \"0efadb03-7701-40f6-b709-67f02db4a787\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.250288 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-catalog-content\") pod \"0efadb03-7701-40f6-b709-67f02db4a787\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.250443 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zb4ss\" (UniqueName: \"kubernetes.io/projected/0efadb03-7701-40f6-b709-67f02db4a787-kube-api-access-zb4ss\") pod \"0efadb03-7701-40f6-b709-67f02db4a787\" (UID: \"0efadb03-7701-40f6-b709-67f02db4a787\") " Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.250828 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-utilities" (OuterVolumeSpecName: "utilities") pod "0efadb03-7701-40f6-b709-67f02db4a787" (UID: "0efadb03-7701-40f6-b709-67f02db4a787"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.258284 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0efadb03-7701-40f6-b709-67f02db4a787-kube-api-access-zb4ss" (OuterVolumeSpecName: "kube-api-access-zb4ss") pod "0efadb03-7701-40f6-b709-67f02db4a787" (UID: "0efadb03-7701-40f6-b709-67f02db4a787"). InnerVolumeSpecName "kube-api-access-zb4ss". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.267319 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "0efadb03-7701-40f6-b709-67f02db4a787" (UID: "0efadb03-7701-40f6-b709-67f02db4a787"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.313656 4956 scope.go:117] "RemoveContainer" containerID="5e847741d07bb4d6f3b750c5a1b08df36f3716a2e8427ce9a73f7a0855c859a0" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.336372 4956 scope.go:117] "RemoveContainer" containerID="038f4844efd8bfd5d6bf9e59b4e217d721c1e513b9fdf871cd8732cd20381aa1" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.352505 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.352546 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/0efadb03-7701-40f6-b709-67f02db4a787-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.352557 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zb4ss\" (UniqueName: \"kubernetes.io/projected/0efadb03-7701-40f6-b709-67f02db4a787-kube-api-access-zb4ss\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.360415 4956 scope.go:117] "RemoveContainer" containerID="f3475b2f7f525a60d6c12ab5fe9923fe6a3a1ef3c573347a59d8b537e5ef224c" Nov 26 17:18:40 crc kubenswrapper[4956]: I1126 17:18:40.387577 4956 scope.go:117] "RemoveContainer" containerID="bb90bf30ca9aa260b87126883ac63132095877dca04b01f21bb7ea444302ab55" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.105916 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-wfbpd" event={"ID":"0efadb03-7701-40f6-b709-67f02db4a787","Type":"ContainerDied","Data":"c6f81de748d01aab61623776a1cb196266b4659c6e8d49884a1fb5b3402a5164"} Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.106022 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-wfbpd" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.106399 4956 scope.go:117] "RemoveContainer" containerID="3cacef257f78e1cc21842ce721426fecad7ed2a4bf87205bf74d14e3027b70df" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.131299 4956 scope.go:117] "RemoveContainer" containerID="c723914df725acdae42f3d1b06550afb389c8da0f39d708158f3d618c8aac28d" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.133583 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfbpd"] Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.141250 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-wfbpd"] Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.151222 4956 scope.go:117] "RemoveContainer" containerID="5441e3554a1c3abe7ede68513aa05752d91de0376ccd8cdbefca62823645d4f0" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.744713 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-db-sync-8hkbb"] Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745237 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="extract-utilities" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745261 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="extract-utilities" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745288 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ee6b955-28a3-45e4-ba60-154ddd97ed1d" containerName="mariadb-database-create" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745296 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ee6b955-28a3-45e4-ba60-154ddd97ed1d" containerName="mariadb-database-create" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745311 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="registry-server" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745319 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="registry-server" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745330 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="registry-server" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745337 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="registry-server" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745351 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="extract-content" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745357 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="extract-content" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745369 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="extract-utilities" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745374 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="extract-utilities" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745390 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ed6c8be2-8540-4937-8ee8-7209c72b1fbe" containerName="mariadb-account-create-update" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745398 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="ed6c8be2-8540-4937-8ee8-7209c72b1fbe" containerName="mariadb-account-create-update" Nov 26 17:18:41 crc kubenswrapper[4956]: E1126 17:18:41.745414 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="extract-content" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745420 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="extract-content" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745558 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="ed6c8be2-8540-4937-8ee8-7209c72b1fbe" containerName="mariadb-account-create-update" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745577 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ee6b955-28a3-45e4-ba60-154ddd97ed1d" containerName="mariadb-database-create" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745588 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="725155ee-d748-4e22-bb8e-8bbde903ecd3" containerName="registry-server" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.745604 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="0efadb03-7701-40f6-b709-67f02db4a787" containerName="registry-server" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.746272 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.749283 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-cgbmz" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.749283 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-config-data" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.755926 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-8hkbb"] Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.880657 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8f65n\" (UniqueName: \"kubernetes.io/projected/eff77660-101d-4609-8ab9-21d93e94567f-kube-api-access-8f65n\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.880763 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-db-sync-config-data\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.880874 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-config-data\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.923223 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzffh"] Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.923627 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-zzffh" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="registry-server" containerID="cri-o://a6b4c9661ffeaae79da3d2b1316bcb9642e1ed29488f45c1fa2293f3aa4922fa" gracePeriod=2 Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.982091 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-db-sync-config-data\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.982191 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-config-data\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.982248 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8f65n\" (UniqueName: \"kubernetes.io/projected/eff77660-101d-4609-8ab9-21d93e94567f-kube-api-access-8f65n\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.989291 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-db-sync-config-data\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:41 crc kubenswrapper[4956]: I1126 17:18:41.989623 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-config-data\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.003938 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8f65n\" (UniqueName: \"kubernetes.io/projected/eff77660-101d-4609-8ab9-21d93e94567f-kube-api-access-8f65n\") pod \"glance-db-sync-8hkbb\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.063779 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.127549 4956 generic.go:334] "Generic (PLEG): container finished" podID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerID="a6b4c9661ffeaae79da3d2b1316bcb9642e1ed29488f45c1fa2293f3aa4922fa" exitCode=0 Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.127609 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzffh" event={"ID":"b4b1de34-cf17-456b-997f-ce8fa99100eb","Type":"ContainerDied","Data":"a6b4c9661ffeaae79da3d2b1316bcb9642e1ed29488f45c1fa2293f3aa4922fa"} Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.406866 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-db-sync-8hkbb"] Nov 26 17:18:42 crc kubenswrapper[4956]: W1126 17:18:42.413490 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeff77660_101d_4609_8ab9_21d93e94567f.slice/crio-c50bc02f42a591fbcfaba7a4c080e3b904ee0654ba1036d87985c13faff8a1c5 WatchSource:0}: Error finding container c50bc02f42a591fbcfaba7a4c080e3b904ee0654ba1036d87985c13faff8a1c5: Status 404 returned error can't find the container with id c50bc02f42a591fbcfaba7a4c080e3b904ee0654ba1036d87985c13faff8a1c5 Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.488149 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.592717 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-utilities\") pod \"b4b1de34-cf17-456b-997f-ce8fa99100eb\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.592981 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-catalog-content\") pod \"b4b1de34-cf17-456b-997f-ce8fa99100eb\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.593014 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fv587\" (UniqueName: \"kubernetes.io/projected/b4b1de34-cf17-456b-997f-ce8fa99100eb-kube-api-access-fv587\") pod \"b4b1de34-cf17-456b-997f-ce8fa99100eb\" (UID: \"b4b1de34-cf17-456b-997f-ce8fa99100eb\") " Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.594067 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-utilities" (OuterVolumeSpecName: "utilities") pod "b4b1de34-cf17-456b-997f-ce8fa99100eb" (UID: "b4b1de34-cf17-456b-997f-ce8fa99100eb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.600139 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b4b1de34-cf17-456b-997f-ce8fa99100eb-kube-api-access-fv587" (OuterVolumeSpecName: "kube-api-access-fv587") pod "b4b1de34-cf17-456b-997f-ce8fa99100eb" (UID: "b4b1de34-cf17-456b-997f-ce8fa99100eb"). InnerVolumeSpecName "kube-api-access-fv587". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.695443 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.695493 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fv587\" (UniqueName: \"kubernetes.io/projected/b4b1de34-cf17-456b-997f-ce8fa99100eb-kube-api-access-fv587\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.724929 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b4b1de34-cf17-456b-997f-ce8fa99100eb" (UID: "b4b1de34-cf17-456b-997f-ce8fa99100eb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:42 crc kubenswrapper[4956]: I1126 17:18:42.797580 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b4b1de34-cf17-456b-997f-ce8fa99100eb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.013279 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0efadb03-7701-40f6-b709-67f02db4a787" path="/var/lib/kubelet/pods/0efadb03-7701-40f6-b709-67f02db4a787/volumes" Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.143116 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8hkbb" event={"ID":"eff77660-101d-4609-8ab9-21d93e94567f","Type":"ContainerStarted","Data":"f7009af7502fd8330d6dc9b66a9eb80360ace39ecf13884756f81912bd97e9ef"} Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.144931 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8hkbb" event={"ID":"eff77660-101d-4609-8ab9-21d93e94567f","Type":"ContainerStarted","Data":"c50bc02f42a591fbcfaba7a4c080e3b904ee0654ba1036d87985c13faff8a1c5"} Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.147261 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-zzffh" event={"ID":"b4b1de34-cf17-456b-997f-ce8fa99100eb","Type":"ContainerDied","Data":"6b0974b1d7aba55052734d842db31878cdd55275f456c42cf54bad96232ff8d4"} Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.147411 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-zzffh" Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.147422 4956 scope.go:117] "RemoveContainer" containerID="a6b4c9661ffeaae79da3d2b1316bcb9642e1ed29488f45c1fa2293f3aa4922fa" Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.173555 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-db-sync-8hkbb" podStartSLOduration=2.173530623 podStartE2EDuration="2.173530623s" podCreationTimestamp="2025-11-26 17:18:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:18:43.16785266 +0000 UTC m=+1568.863813252" watchObservedRunningTime="2025-11-26 17:18:43.173530623 +0000 UTC m=+1568.869491185" Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.193624 4956 scope.go:117] "RemoveContainer" containerID="39293835d9adeb89ba8bdd70c03a5f7b4716af5e8782eb9c4fa7dc36d465ad0f" Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.196464 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-zzffh"] Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.245923 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-zzffh"] Nov 26 17:18:43 crc kubenswrapper[4956]: I1126 17:18:43.258434 4956 scope.go:117] "RemoveContainer" containerID="db42b20f349feaf558d74ea4e214636d5ca37c9b2ec784fc4303beae4b9f4be4" Nov 26 17:18:45 crc kubenswrapper[4956]: I1126 17:18:45.014149 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" path="/var/lib/kubelet/pods/b4b1de34-cf17-456b-997f-ce8fa99100eb/volumes" Nov 26 17:18:46 crc kubenswrapper[4956]: I1126 17:18:46.179941 4956 generic.go:334] "Generic (PLEG): container finished" podID="eff77660-101d-4609-8ab9-21d93e94567f" containerID="f7009af7502fd8330d6dc9b66a9eb80360ace39ecf13884756f81912bd97e9ef" exitCode=0 Nov 26 17:18:46 crc kubenswrapper[4956]: I1126 17:18:46.180022 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8hkbb" event={"ID":"eff77660-101d-4609-8ab9-21d93e94567f","Type":"ContainerDied","Data":"f7009af7502fd8330d6dc9b66a9eb80360ace39ecf13884756f81912bd97e9ef"} Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.547200 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.723187 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-config-data\") pod \"eff77660-101d-4609-8ab9-21d93e94567f\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.723804 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8f65n\" (UniqueName: \"kubernetes.io/projected/eff77660-101d-4609-8ab9-21d93e94567f-kube-api-access-8f65n\") pod \"eff77660-101d-4609-8ab9-21d93e94567f\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.723982 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-db-sync-config-data\") pod \"eff77660-101d-4609-8ab9-21d93e94567f\" (UID: \"eff77660-101d-4609-8ab9-21d93e94567f\") " Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.744723 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eff77660-101d-4609-8ab9-21d93e94567f-kube-api-access-8f65n" (OuterVolumeSpecName: "kube-api-access-8f65n") pod "eff77660-101d-4609-8ab9-21d93e94567f" (UID: "eff77660-101d-4609-8ab9-21d93e94567f"). InnerVolumeSpecName "kube-api-access-8f65n". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.745058 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "eff77660-101d-4609-8ab9-21d93e94567f" (UID: "eff77660-101d-4609-8ab9-21d93e94567f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.772044 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-config-data" (OuterVolumeSpecName: "config-data") pod "eff77660-101d-4609-8ab9-21d93e94567f" (UID: "eff77660-101d-4609-8ab9-21d93e94567f"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.825986 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.826049 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8f65n\" (UniqueName: \"kubernetes.io/projected/eff77660-101d-4609-8ab9-21d93e94567f-kube-api-access-8f65n\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:47 crc kubenswrapper[4956]: I1126 17:18:47.826068 4956 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/eff77660-101d-4609-8ab9-21d93e94567f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:48 crc kubenswrapper[4956]: I1126 17:18:48.205919 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-db-sync-8hkbb" event={"ID":"eff77660-101d-4609-8ab9-21d93e94567f","Type":"ContainerDied","Data":"c50bc02f42a591fbcfaba7a4c080e3b904ee0654ba1036d87985c13faff8a1c5"} Nov 26 17:18:48 crc kubenswrapper[4956]: I1126 17:18:48.205994 4956 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c50bc02f42a591fbcfaba7a4c080e3b904ee0654ba1036d87985c13faff8a1c5" Nov 26 17:18:48 crc kubenswrapper[4956]: I1126 17:18:48.206005 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-db-sync-8hkbb" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.471705 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:49 crc kubenswrapper[4956]: E1126 17:18:49.472589 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="registry-server" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.472607 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="registry-server" Nov 26 17:18:49 crc kubenswrapper[4956]: E1126 17:18:49.472667 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="extract-utilities" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.472677 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="extract-utilities" Nov 26 17:18:49 crc kubenswrapper[4956]: E1126 17:18:49.472696 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eff77660-101d-4609-8ab9-21d93e94567f" containerName="glance-db-sync" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.472704 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="eff77660-101d-4609-8ab9-21d93e94567f" containerName="glance-db-sync" Nov 26 17:18:49 crc kubenswrapper[4956]: E1126 17:18:49.472736 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="extract-content" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.472742 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="extract-content" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.472924 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b4b1de34-cf17-456b-997f-ce8fa99100eb" containerName="registry-server" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.472973 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="eff77660-101d-4609-8ab9-21d93e94567f" containerName="glance-db-sync" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.473924 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.476441 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-default-single-config-data" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.476464 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-glance-dockercfg-cgbmz" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.477019 4956 reflector.go:368] Caches populated for *v1.Secret from object-"glance-kuttl-tests"/"glance-scripts" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.521797 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.550707 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.552309 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.567473 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.656847 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.656923 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.656985 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzpqn\" (UniqueName: \"kubernetes.io/projected/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-kube-api-access-jzpqn\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657015 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-config-data\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657041 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-scripts\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657160 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-scripts\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657227 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657289 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-lib-modules\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657323 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-dev\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657376 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-lib-modules\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657409 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657456 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-logs\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657552 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-httpd-run\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657603 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657635 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657672 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-dev\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657717 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657742 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-logs\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657783 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-run\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657820 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-sys\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657855 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-sys\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657901 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-nvme\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657957 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.657985 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-nmxml\" (UniqueName: \"kubernetes.io/projected/b548ac15-9451-4ac9-b353-503c528c69de-kube-api-access-nmxml\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.658014 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-httpd-run\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.658049 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.658118 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-run\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.658148 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-config-data\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.759830 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-httpd-run\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.759915 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.759944 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.759969 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-dev\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.759993 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760010 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-logs\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760033 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-run\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760057 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-sys\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760079 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-sys\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760097 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-nvme\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760130 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760157 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-nmxml\" (UniqueName: \"kubernetes.io/projected/b548ac15-9451-4ac9-b353-503c528c69de-kube-api-access-nmxml\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760188 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-httpd-run\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760216 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760252 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-run\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760273 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-config-data\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760310 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760331 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760397 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzpqn\" (UniqueName: \"kubernetes.io/projected/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-kube-api-access-jzpqn\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760430 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-config-data\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760460 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-scripts\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760493 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-scripts\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760515 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760544 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-lib-modules\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760577 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-dev\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760804 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-lib-modules\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760826 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.760846 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-logs\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.761452 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-logs\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.761822 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-httpd-run\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.761884 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.762268 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.762946 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-run\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763029 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-dev\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763079 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-etc-nvme\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763431 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-logs\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763478 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-run\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763505 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-sys\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763534 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-sys\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763580 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-nvme\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.763613 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-etc-iscsi\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.764238 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-httpd-run\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.764347 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") device mount path \"/mnt/openstack/pv06\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.765202 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.765371 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") device mount path \"/mnt/openstack/pv19\"" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.766388 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-lib-modules\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.766553 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.766616 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-lib-modules\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.766631 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-var-locks-brick\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.766672 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-dev\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.778319 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-config-data\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.778348 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-config-data\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.780653 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-scripts\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.785209 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-scripts\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.789956 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzpqn\" (UniqueName: \"kubernetes.io/projected/d6f63b3b-78e6-46cc-ab34-685c5deba6f4-kube-api-access-jzpqn\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.789971 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-nmxml\" (UniqueName: \"kubernetes.io/projected/b548ac15-9451-4ac9-b353-503c528c69de-kube-api-access-nmxml\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.792817 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.793062 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.802140 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage19-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage19-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.812472 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"glance-default-single-1\" (UID: \"d6f63b3b-78e6-46cc-ab34-685c5deba6f4\") " pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:49 crc kubenswrapper[4956]: I1126 17:18:49.872654 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:50 crc kubenswrapper[4956]: I1126 17:18:50.093583 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:18:50 crc kubenswrapper[4956]: I1126 17:18:50.385077 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:18:50 crc kubenswrapper[4956]: I1126 17:18:50.537425 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-1"] Nov 26 17:18:50 crc kubenswrapper[4956]: W1126 17:18:50.540303 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd6f63b3b_78e6_46cc_ab34_685c5deba6f4.slice/crio-92cab6c7224eec427f63eec6ac75f280049daaf63edd7671b17d84e7546699f9 WatchSource:0}: Error finding container 92cab6c7224eec427f63eec6ac75f280049daaf63edd7671b17d84e7546699f9: Status 404 returned error can't find the container with id 92cab6c7224eec427f63eec6ac75f280049daaf63edd7671b17d84e7546699f9 Nov 26 17:18:50 crc kubenswrapper[4956]: I1126 17:18:50.931469 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2slwb"] Nov 26 17:18:50 crc kubenswrapper[4956]: I1126 17:18:50.932819 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-2slwb" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="registry-server" containerID="cri-o://6000a717f570741b549c70043af28fd1812f02a54b710fddf79b2c13e6f1f8e6" gracePeriod=2 Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.235016 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d6f63b3b-78e6-46cc-ab34-685c5deba6f4","Type":"ContainerStarted","Data":"a6fae633748c5d858801394cee876e7f4e379e3862894f0a285f915a83a5294a"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.235645 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d6f63b3b-78e6-46cc-ab34-685c5deba6f4","Type":"ContainerStarted","Data":"7b308c6edde906a3b547238c2bfe711f0105eae51221d2eb08fceda4468b2284"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.235668 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-1" event={"ID":"d6f63b3b-78e6-46cc-ab34-685c5deba6f4","Type":"ContainerStarted","Data":"92cab6c7224eec427f63eec6ac75f280049daaf63edd7671b17d84e7546699f9"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.239194 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerDied","Data":"6000a717f570741b549c70043af28fd1812f02a54b710fddf79b2c13e6f1f8e6"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.239374 4956 generic.go:334] "Generic (PLEG): container finished" podID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerID="6000a717f570741b549c70043af28fd1812f02a54b710fddf79b2c13e6f1f8e6" exitCode=0 Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.271562 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b548ac15-9451-4ac9-b353-503c528c69de","Type":"ContainerStarted","Data":"4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.272158 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b548ac15-9451-4ac9-b353-503c528c69de","Type":"ContainerStarted","Data":"fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.272177 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b548ac15-9451-4ac9-b353-503c528c69de","Type":"ContainerStarted","Data":"f57a5e095229e48727dfaec9f66b828016e66e54c519f2c22aff4f4b5785971d"} Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.299420 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-1" podStartSLOduration=2.299394173 podStartE2EDuration="2.299394173s" podCreationTimestamp="2025-11-26 17:18:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:18:51.269058748 +0000 UTC m=+1576.965019300" watchObservedRunningTime="2025-11-26 17:18:51.299394173 +0000 UTC m=+1576.995354725" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.304468 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=3.304439118 podStartE2EDuration="3.304439118s" podCreationTimestamp="2025-11-26 17:18:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:18:51.290207608 +0000 UTC m=+1576.986168160" watchObservedRunningTime="2025-11-26 17:18:51.304439118 +0000 UTC m=+1577.000399680" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.494691 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.589846 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9h5wc\" (UniqueName: \"kubernetes.io/projected/47ebddca-2799-4e01-9b63-a62a2f0398fb-kube-api-access-9h5wc\") pod \"47ebddca-2799-4e01-9b63-a62a2f0398fb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.589997 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-utilities\") pod \"47ebddca-2799-4e01-9b63-a62a2f0398fb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.590116 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-catalog-content\") pod \"47ebddca-2799-4e01-9b63-a62a2f0398fb\" (UID: \"47ebddca-2799-4e01-9b63-a62a2f0398fb\") " Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.591515 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-utilities" (OuterVolumeSpecName: "utilities") pod "47ebddca-2799-4e01-9b63-a62a2f0398fb" (UID: "47ebddca-2799-4e01-9b63-a62a2f0398fb"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.596761 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/47ebddca-2799-4e01-9b63-a62a2f0398fb-kube-api-access-9h5wc" (OuterVolumeSpecName: "kube-api-access-9h5wc") pod "47ebddca-2799-4e01-9b63-a62a2f0398fb" (UID: "47ebddca-2799-4e01-9b63-a62a2f0398fb"). InnerVolumeSpecName "kube-api-access-9h5wc". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.655137 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "47ebddca-2799-4e01-9b63-a62a2f0398fb" (UID: "47ebddca-2799-4e01-9b63-a62a2f0398fb"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.692529 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9h5wc\" (UniqueName: \"kubernetes.io/projected/47ebddca-2799-4e01-9b63-a62a2f0398fb-kube-api-access-9h5wc\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.692575 4956 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-utilities\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:51 crc kubenswrapper[4956]: I1126 17:18:51.692589 4956 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/47ebddca-2799-4e01-9b63-a62a2f0398fb-catalog-content\") on node \"crc\" DevicePath \"\"" Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.285266 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-2slwb" event={"ID":"47ebddca-2799-4e01-9b63-a62a2f0398fb","Type":"ContainerDied","Data":"50aac5131262b31428bd79b476aab10435770d7b41eb9d943af97893cd0cee0c"} Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.286048 4956 scope.go:117] "RemoveContainer" containerID="6000a717f570741b549c70043af28fd1812f02a54b710fddf79b2c13e6f1f8e6" Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.285362 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-2slwb" Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.319176 4956 scope.go:117] "RemoveContainer" containerID="cc3501b013264bcd5d26ae9c5d93157b35693fcee9a6705695c27c83188f4094" Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.324341 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-2slwb"] Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.331833 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-2slwb"] Nov 26 17:18:52 crc kubenswrapper[4956]: I1126 17:18:52.355007 4956 scope.go:117] "RemoveContainer" containerID="b2b23a47b9a2c4b3048cb885ea355114bf927f2af4a49ba936b57eea6ae9cce5" Nov 26 17:18:53 crc kubenswrapper[4956]: I1126 17:18:53.009697 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" path="/var/lib/kubelet/pods/47ebddca-2799-4e01-9b63-a62a2f0398fb/volumes" Nov 26 17:18:53 crc kubenswrapper[4956]: I1126 17:18:53.996186 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:18:53 crc kubenswrapper[4956]: E1126 17:18:53.996611 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:18:59 crc kubenswrapper[4956]: I1126 17:18:59.051202 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-create-76k8h"] Nov 26 17:18:59 crc kubenswrapper[4956]: I1126 17:18:59.057825 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-create-76k8h"] Nov 26 17:18:59 crc kubenswrapper[4956]: I1126 17:18:59.873744 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:59 crc kubenswrapper[4956]: I1126 17:18:59.874088 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:59 crc kubenswrapper[4956]: I1126 17:18:59.910566 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:18:59 crc kubenswrapper[4956]: I1126 17:18:59.931884 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.034618 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb"] Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.045994 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-8bae-account-create-update-8x9jb"] Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.094789 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.095562 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.126198 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.132161 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.376503 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.376606 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.376624 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:00 crc kubenswrapper[4956]: I1126 17:19:00.376640 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:01 crc kubenswrapper[4956]: I1126 17:19:01.005347 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a79036d2-2c7e-4b04-a1bc-b194074b27e7" path="/var/lib/kubelet/pods/a79036d2-2c7e-4b04-a1bc-b194074b27e7/volumes" Nov 26 17:19:01 crc kubenswrapper[4956]: I1126 17:19:01.006129 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="be4a1c0e-bb65-4419-9349-708ffe694948" path="/var/lib/kubelet/pods/be4a1c0e-bb65-4419-9349-708ffe694948/volumes" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.392882 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.393177 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.698337 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.698526 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.714090 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-1" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.731388 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.766827 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:02 crc kubenswrapper[4956]: I1126 17:19:02.786982 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:19:04 crc kubenswrapper[4956]: I1126 17:19:04.414705 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-log" containerID="cri-o://fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd" gracePeriod=30 Nov 26 17:19:04 crc kubenswrapper[4956]: I1126 17:19:04.414854 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="glance-kuttl-tests/glance-default-single-0" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-httpd" containerID="cri-o://4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce" gracePeriod=30 Nov 26 17:19:04 crc kubenswrapper[4956]: I1126 17:19:04.425334 4956 prober.go:107] "Probe failed" probeType="Readiness" pod="glance-kuttl-tests/glance-default-single-0" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-httpd" probeResult="failure" output="Get \"http://10.217.0.155:9292/healthcheck\": EOF" Nov 26 17:19:05 crc kubenswrapper[4956]: I1126 17:19:05.427669 4956 generic.go:334] "Generic (PLEG): container finished" podID="b548ac15-9451-4ac9-b353-503c528c69de" containerID="fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd" exitCode=143 Nov 26 17:19:05 crc kubenswrapper[4956]: I1126 17:19:05.427754 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b548ac15-9451-4ac9-b353-503c528c69de","Type":"ContainerDied","Data":"fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd"} Nov 26 17:19:07 crc kubenswrapper[4956]: I1126 17:19:07.995672 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:19:07 crc kubenswrapper[4956]: E1126 17:19:07.996398 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.361756 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426466 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-logs\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426536 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-iscsi\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426618 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance-cache\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426659 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-config-data\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426706 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-dev\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426743 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-scripts\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426778 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-run\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426829 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-nvme\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426850 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-sys\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426893 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426920 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-var-locks-brick\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426961 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-lib-modules\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.426997 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-httpd-run\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.427031 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nmxml\" (UniqueName: \"kubernetes.io/projected/b548ac15-9451-4ac9-b353-503c528c69de-kube-api-access-nmxml\") pod \"b548ac15-9451-4ac9-b353-503c528c69de\" (UID: \"b548ac15-9451-4ac9-b353-503c528c69de\") " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.427208 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-run" (OuterVolumeSpecName: "run") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.427609 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-lib-modules" (OuterVolumeSpecName: "lib-modules") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "lib-modules". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.427687 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-var-locks-brick" (OuterVolumeSpecName: "var-locks-brick") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "var-locks-brick". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.427701 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-dev" (OuterVolumeSpecName: "dev") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "dev". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428082 4956 reconciler_common.go:293] "Volume detached for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-dev\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428108 4956 reconciler_common.go:293] "Volume detached for volume \"run\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428130 4956 reconciler_common.go:293] "Volume detached for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-var-locks-brick\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428149 4956 reconciler_common.go:293] "Volume detached for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-lib-modules\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.427608 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-iscsi" (OuterVolumeSpecName: "etc-iscsi") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "etc-iscsi". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428230 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-nvme" (OuterVolumeSpecName: "etc-nvme") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "etc-nvme". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428276 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-sys" (OuterVolumeSpecName: "sys") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "sys". PluginName "kubernetes.io/host-path", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428628 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-logs" (OuterVolumeSpecName: "logs") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.428838 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.434141 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage17-crc" (OuterVolumeSpecName: "glance") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "local-storage17-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.434784 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b548ac15-9451-4ac9-b353-503c528c69de-kube-api-access-nmxml" (OuterVolumeSpecName: "kube-api-access-nmxml") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "kube-api-access-nmxml". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.443189 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-scripts" (OuterVolumeSpecName: "scripts") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.456646 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage16-crc" (OuterVolumeSpecName: "glance-cache") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "local-storage16-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.487428 4956 generic.go:334] "Generic (PLEG): container finished" podID="b548ac15-9451-4ac9-b353-503c528c69de" containerID="4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce" exitCode=0 Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.487551 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.487541 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b548ac15-9451-4ac9-b353-503c528c69de","Type":"ContainerDied","Data":"4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce"} Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.487888 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"b548ac15-9451-4ac9-b353-503c528c69de","Type":"ContainerDied","Data":"f57a5e095229e48727dfaec9f66b828016e66e54c519f2c22aff4f4b5785971d"} Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.487939 4956 scope.go:117] "RemoveContainer" containerID="4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.499595 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-config-data" (OuterVolumeSpecName: "config-data") pod "b548ac15-9451-4ac9-b353-503c528c69de" (UID: "b548ac15-9451-4ac9-b353-503c528c69de"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.516910 4956 scope.go:117] "RemoveContainer" containerID="fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530571 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530618 4956 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-config-data\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530634 4956 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b548ac15-9451-4ac9-b353-503c528c69de-scripts\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530647 4956 reconciler_common.go:293] "Volume detached for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-nvme\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530661 4956 reconciler_common.go:293] "Volume detached for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-sys\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530679 4956 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" " Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530690 4956 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-httpd-run\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530705 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nmxml\" (UniqueName: \"kubernetes.io/projected/b548ac15-9451-4ac9-b353-503c528c69de-kube-api-access-nmxml\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530713 4956 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b548ac15-9451-4ac9-b353-503c528c69de-logs\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.530721 4956 reconciler_common.go:293] "Volume detached for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/b548ac15-9451-4ac9-b353-503c528c69de-etc-iscsi\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.542836 4956 scope.go:117] "RemoveContainer" containerID="4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce" Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.543658 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce\": container with ID starting with 4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce not found: ID does not exist" containerID="4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.543809 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce"} err="failed to get container status \"4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce\": rpc error: code = NotFound desc = could not find container \"4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce\": container with ID starting with 4f6d7cc4c7d92a20409a1cfbd51eed12d5e78b2ec7f5d117487c296cf43210ce not found: ID does not exist" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.543945 4956 scope.go:117] "RemoveContainer" containerID="fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd" Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.544437 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd\": container with ID starting with fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd not found: ID does not exist" containerID="fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.544550 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd"} err="failed to get container status \"fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd\": rpc error: code = NotFound desc = could not find container \"fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd\": container with ID starting with fd302cb643ab92bed2bbfdde59cd21f8d566536c5797634db77dd3cb2a8babbd not found: ID does not exist" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.547388 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage17-crc" (UniqueName: "kubernetes.io/local-volume/local-storage17-crc") on node "crc" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.547640 4956 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage16-crc" (UniqueName: "kubernetes.io/local-volume/local-storage16-crc") on node "crc" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.632454 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.633100 4956 reconciler_common.go:293] "Volume detached for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") on node \"crc\" DevicePath \"\"" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.836503 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.846886 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.859507 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.862596 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="registry-server" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862634 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="registry-server" Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.862650 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="extract-utilities" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862660 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="extract-utilities" Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.862693 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-log" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862700 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-log" Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.862717 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="extract-content" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862724 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="extract-content" Nov 26 17:19:09 crc kubenswrapper[4956]: E1126 17:19:09.862734 4956 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-httpd" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862740 4956 state_mem.go:107] "Deleted CPUSet assignment" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-httpd" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862974 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-httpd" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.862996 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="b548ac15-9451-4ac9-b353-503c528c69de" containerName="glance-log" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.863010 4956 memory_manager.go:354] "RemoveStaleState removing state" podUID="47ebddca-2799-4e01-9b63-a62a2f0398fb" containerName="registry-server" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.864013 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.875120 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938356 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-run\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938425 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d01d93-2c13-4500-ad93-e82c70e3df5d-logs\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938457 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-lib-modules\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938488 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938520 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938543 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-dev\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938573 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m8wkx\" (UniqueName: \"kubernetes.io/projected/99d01d93-2c13-4500-ad93-e82c70e3df5d-kube-api-access-m8wkx\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938614 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d01d93-2c13-4500-ad93-e82c70e3df5d-scripts\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.938710 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.939057 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-etc-nvme\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.939146 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.939212 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99d01d93-2c13-4500-ad93-e82c70e3df5d-httpd-run\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.939312 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d01d93-2c13-4500-ad93-e82c70e3df5d-config-data\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:09 crc kubenswrapper[4956]: I1126 17:19:09.939353 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-sys\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041038 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041132 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041162 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-dev\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041201 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m8wkx\" (UniqueName: \"kubernetes.io/projected/99d01d93-2c13-4500-ad93-e82c70e3df5d-kube-api-access-m8wkx\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041260 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d01d93-2c13-4500-ad93-e82c70e3df5d-scripts\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041295 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041356 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-etc-nvme\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041379 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99d01d93-2c13-4500-ad93-e82c70e3df5d-httpd-run\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041402 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041440 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d01d93-2c13-4500-ad93-e82c70e3df5d-config-data\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041467 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-sys\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041581 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-run\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041629 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d01d93-2c13-4500-ad93-e82c70e3df5d-logs\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041665 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-lib-modules\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041758 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lib-modules\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-lib-modules\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.041807 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-iscsi\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-etc-iscsi\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.042784 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-run\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.042793 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") device mount path \"/mnt/openstack/pv16\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043164 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/99d01d93-2c13-4500-ad93-e82c70e3df5d-logs\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043318 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-nvme\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-etc-nvme\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043647 4956 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") device mount path \"/mnt/openstack/pv17\"" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043813 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/99d01d93-2c13-4500-ad93-e82c70e3df5d-httpd-run\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043844 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dev\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-dev\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043821 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-locks-brick\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-var-locks-brick\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.043950 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sys\" (UniqueName: \"kubernetes.io/host-path/99d01d93-2c13-4500-ad93-e82c70e3df5d-sys\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.050576 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/99d01d93-2c13-4500-ad93-e82c70e3df5d-scripts\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.052443 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/99d01d93-2c13-4500-ad93-e82c70e3df5d-config-data\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.069117 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage16-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage16-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.074893 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m8wkx\" (UniqueName: \"kubernetes.io/projected/99d01d93-2c13-4500-ad93-e82c70e3df5d-kube-api-access-m8wkx\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.089473 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage17-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage17-crc\") pod \"glance-default-single-0\" (UID: \"99d01d93-2c13-4500-ad93-e82c70e3df5d\") " pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.198387 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:10 crc kubenswrapper[4956]: W1126 17:19:10.705033 4956 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod99d01d93_2c13_4500_ad93_e82c70e3df5d.slice/crio-1b774dd893007ce962b6eb6b6d8044512d84cb70c59dfa5f1082873f2a85341d WatchSource:0}: Error finding container 1b774dd893007ce962b6eb6b6d8044512d84cb70c59dfa5f1082873f2a85341d: Status 404 returned error can't find the container with id 1b774dd893007ce962b6eb6b6d8044512d84cb70c59dfa5f1082873f2a85341d Nov 26 17:19:10 crc kubenswrapper[4956]: I1126 17:19:10.707689 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["glance-kuttl-tests/glance-default-single-0"] Nov 26 17:19:11 crc kubenswrapper[4956]: I1126 17:19:11.010325 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b548ac15-9451-4ac9-b353-503c528c69de" path="/var/lib/kubelet/pods/b548ac15-9451-4ac9-b353-503c528c69de/volumes" Nov 26 17:19:11 crc kubenswrapper[4956]: I1126 17:19:11.515249 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"99d01d93-2c13-4500-ad93-e82c70e3df5d","Type":"ContainerStarted","Data":"07e16c772116e391d8fb831616867205a69bd63a3c920860d367479fbbdd6ede"} Nov 26 17:19:11 crc kubenswrapper[4956]: I1126 17:19:11.515928 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"99d01d93-2c13-4500-ad93-e82c70e3df5d","Type":"ContainerStarted","Data":"b483d6897d5358fb977f2c7db184195ed167000f9e08b99a0322cf206e3f513f"} Nov 26 17:19:11 crc kubenswrapper[4956]: I1126 17:19:11.515956 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="glance-kuttl-tests/glance-default-single-0" event={"ID":"99d01d93-2c13-4500-ad93-e82c70e3df5d","Type":"ContainerStarted","Data":"1b774dd893007ce962b6eb6b6d8044512d84cb70c59dfa5f1082873f2a85341d"} Nov 26 17:19:11 crc kubenswrapper[4956]: I1126 17:19:11.554421 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="glance-kuttl-tests/glance-default-single-0" podStartSLOduration=2.554383883 podStartE2EDuration="2.554383883s" podCreationTimestamp="2025-11-26 17:19:09 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-11-26 17:19:11.539423912 +0000 UTC m=+1597.235384544" watchObservedRunningTime="2025-11-26 17:19:11.554383883 +0000 UTC m=+1597.250344475" Nov 26 17:19:19 crc kubenswrapper[4956]: I1126 17:19:19.996454 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:19:19 crc kubenswrapper[4956]: E1126 17:19:19.997427 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:19:20 crc kubenswrapper[4956]: I1126 17:19:20.200202 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:20 crc kubenswrapper[4956]: I1126 17:19:20.200289 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:20 crc kubenswrapper[4956]: I1126 17:19:20.236968 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:20 crc kubenswrapper[4956]: I1126 17:19:20.256152 4956 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:20 crc kubenswrapper[4956]: I1126 17:19:20.644937 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:20 crc kubenswrapper[4956]: I1126 17:19:20.644994 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:22 crc kubenswrapper[4956]: I1126 17:19:22.816435 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:22 crc kubenswrapper[4956]: I1126 17:19:22.817372 4956 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Nov 26 17:19:22 crc kubenswrapper[4956]: I1126 17:19:22.823481 4956 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="glance-kuttl-tests/glance-default-single-0" Nov 26 17:19:26 crc kubenswrapper[4956]: I1126 17:19:26.056332 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7866q"] Nov 26 17:19:26 crc kubenswrapper[4956]: I1126 17:19:26.065082 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-db-sync-7866q"] Nov 26 17:19:27 crc kubenswrapper[4956]: I1126 17:19:27.016951 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8d16e0ee-c3be-4697-a639-a3274721d828" path="/var/lib/kubelet/pods/8d16e0ee-c3be-4697-a639-a3274721d828/volumes" Nov 26 17:19:31 crc kubenswrapper[4956]: I1126 17:19:31.995652 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:19:31 crc kubenswrapper[4956]: E1126 17:19:31.997001 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:19:33 crc kubenswrapper[4956]: I1126 17:19:33.036928 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6txpz"] Nov 26 17:19:33 crc kubenswrapper[4956]: I1126 17:19:33.048598 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["glance-kuttl-tests/keystone-bootstrap-6txpz"] Nov 26 17:19:35 crc kubenswrapper[4956]: I1126 17:19:35.010941 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a096a446-ae0a-4e84-88a6-2fe7d8ca35f0" path="/var/lib/kubelet/pods/a096a446-ae0a-4e84-88a6-2fe7d8ca35f0/volumes" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.613400 4956 scope.go:117] "RemoveContainer" containerID="aeea6d747ce568ce405549566998493dff831b6fbff9f50c80d92e7898d0a13c" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.647410 4956 scope.go:117] "RemoveContainer" containerID="daca6791b8b1af0594e0b98f546c98e3dca5719bb46700a72aa7afd7b567cd8f" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.703045 4956 scope.go:117] "RemoveContainer" containerID="648e44c093b0e44c9efb0fcc198e73caa1ef6477ce350814cc0a7a8dc82621c0" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.748452 4956 scope.go:117] "RemoveContainer" containerID="b49fad21e2c286eb9f874fffb6845cd8208944ff5635e0a6e0e7fb4f4ec244eb" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.779647 4956 scope.go:117] "RemoveContainer" containerID="d7219e7cc62d280e0b3569ebf752e6643236c7aed32e34571bee8d4f49a9ada9" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.837239 4956 scope.go:117] "RemoveContainer" containerID="18d72b5b430422cb8a3b35cc20b68088ca84c2df2e30d9e4428c2ce987544c09" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.904183 4956 scope.go:117] "RemoveContainer" containerID="0810e3706926c9d3f1498390eb1e795e8d17c4af87fa04dbdc4932d5ff1ca150" Nov 26 17:19:40 crc kubenswrapper[4956]: I1126 17:19:40.932007 4956 scope.go:117] "RemoveContainer" containerID="62e61e25806244bd7858bc61f6f04696cce99aa4c947ba0fda6486fc883f4ed0" Nov 26 17:19:46 crc kubenswrapper[4956]: I1126 17:19:46.995745 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:19:46 crc kubenswrapper[4956]: E1126 17:19:46.996814 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.458868 4956 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-9l4hp/must-gather-92ljh"] Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.461068 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.464058 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9l4hp"/"openshift-service-ca.crt" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.464114 4956 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-9l4hp"/"kube-root-ca.crt" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.464275 4956 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-9l4hp"/"default-dockercfg-8kszh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.472388 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9l4hp/must-gather-92ljh"] Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.493535 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-79mq7\" (UniqueName: \"kubernetes.io/projected/567fade8-9828-483f-b993-261298068736-kube-api-access-79mq7\") pod \"must-gather-92ljh\" (UID: \"567fade8-9828-483f-b993-261298068736\") " pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.493694 4956 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/567fade8-9828-483f-b993-261298068736-must-gather-output\") pod \"must-gather-92ljh\" (UID: \"567fade8-9828-483f-b993-261298068736\") " pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.594905 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/567fade8-9828-483f-b993-261298068736-must-gather-output\") pod \"must-gather-92ljh\" (UID: \"567fade8-9828-483f-b993-261298068736\") " pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.594965 4956 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-79mq7\" (UniqueName: \"kubernetes.io/projected/567fade8-9828-483f-b993-261298068736-kube-api-access-79mq7\") pod \"must-gather-92ljh\" (UID: \"567fade8-9828-483f-b993-261298068736\") " pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.595695 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/567fade8-9828-483f-b993-261298068736-must-gather-output\") pod \"must-gather-92ljh\" (UID: \"567fade8-9828-483f-b993-261298068736\") " pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.618835 4956 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-79mq7\" (UniqueName: \"kubernetes.io/projected/567fade8-9828-483f-b993-261298068736-kube-api-access-79mq7\") pod \"must-gather-92ljh\" (UID: \"567fade8-9828-483f-b993-261298068736\") " pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:57 crc kubenswrapper[4956]: I1126 17:19:57.779329 4956 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:19:58 crc kubenswrapper[4956]: I1126 17:19:58.198800 4956 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-9l4hp/must-gather-92ljh"] Nov 26 17:19:58 crc kubenswrapper[4956]: I1126 17:19:58.210072 4956 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Nov 26 17:19:59 crc kubenswrapper[4956]: I1126 17:19:59.079108 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9l4hp/must-gather-92ljh" event={"ID":"567fade8-9828-483f-b993-261298068736","Type":"ContainerStarted","Data":"94465eec531e12ee66e23b1560d4501123ccf06b1d46a346b710fb054c5b8e9a"} Nov 26 17:19:59 crc kubenswrapper[4956]: I1126 17:19:59.995949 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:19:59 crc kubenswrapper[4956]: E1126 17:19:59.996238 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:20:03 crc kubenswrapper[4956]: I1126 17:20:03.121470 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9l4hp/must-gather-92ljh" event={"ID":"567fade8-9828-483f-b993-261298068736","Type":"ContainerStarted","Data":"5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa"} Nov 26 17:20:03 crc kubenswrapper[4956]: I1126 17:20:03.123962 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9l4hp/must-gather-92ljh" event={"ID":"567fade8-9828-483f-b993-261298068736","Type":"ContainerStarted","Data":"b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9"} Nov 26 17:20:03 crc kubenswrapper[4956]: I1126 17:20:03.147928 4956 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-9l4hp/must-gather-92ljh" podStartSLOduration=2.250912325 podStartE2EDuration="6.147903551s" podCreationTimestamp="2025-11-26 17:19:57 +0000 UTC" firstStartedPulling="2025-11-26 17:19:58.209806726 +0000 UTC m=+1643.905767278" lastFinishedPulling="2025-11-26 17:20:02.106797952 +0000 UTC m=+1647.802758504" observedRunningTime="2025-11-26 17:20:03.142170156 +0000 UTC m=+1648.838130708" watchObservedRunningTime="2025-11-26 17:20:03.147903551 +0000 UTC m=+1648.843864123" Nov 26 17:20:11 crc kubenswrapper[4956]: I1126 17:20:11.995619 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:20:11 crc kubenswrapper[4956]: E1126 17:20:11.996708 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:20:26 crc kubenswrapper[4956]: I1126 17:20:26.996782 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:20:26 crc kubenswrapper[4956]: E1126 17:20:26.998382 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:20:41 crc kubenswrapper[4956]: I1126 17:20:41.140510 4956 scope.go:117] "RemoveContainer" containerID="913d2b9945bf65d004afee91480a370edf66ba6fd49aeb48ed80fbb0a58233d6" Nov 26 17:20:41 crc kubenswrapper[4956]: I1126 17:20:41.191904 4956 scope.go:117] "RemoveContainer" containerID="7ad72aba18c790174f8236ebb4ed75f6dab833a1d5d7b3d3ab22b3829ed32d0e" Nov 26 17:20:41 crc kubenswrapper[4956]: I1126 17:20:41.209300 4956 scope.go:117] "RemoveContainer" containerID="0a27042a4bfa3e06491627e23c259bd6a35d0930c4424e580ce4f5f5c7358f93" Nov 26 17:20:41 crc kubenswrapper[4956]: I1126 17:20:41.239930 4956 scope.go:117] "RemoveContainer" containerID="bc5b585a847967f69417407326974c1faa7005b635172202cd81ea67765ccd00" Nov 26 17:20:41 crc kubenswrapper[4956]: I1126 17:20:41.953509 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/util/0.log" Nov 26 17:20:41 crc kubenswrapper[4956]: I1126 17:20:41.995937 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:20:41 crc kubenswrapper[4956]: E1126 17:20:41.996611 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.158584 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/util/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.177150 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/pull/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.229755 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/pull/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.380189 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/extract/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.381434 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/util/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.439282 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_10ada0794789fef111fe8e7650cd7f9005cab5cb91c24f9efb71290030mmz8l_c6bd841b-0f31-4fd1-94c2-d776ffe4582c/pull/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.607591 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/util/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.817081 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/util/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.822637 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/pull/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.822753 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/pull/0.log" Nov 26 17:20:42 crc kubenswrapper[4956]: I1126 17:20:42.995472 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/extract/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.028665 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/util/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.034780 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_27e8bc079695f3aed52a6c5be68196d91a6230a1a03a8fc87a19aa534fvrxw2_172f2407-1798-487a-b864-27f28d867ecd/pull/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.216077 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/util/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.370137 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/util/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.391759 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/pull/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.424682 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/pull/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.619407 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/pull/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.623161 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/extract/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.652343 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_5d473c3169f40b179d14921c90af2c8546b7b757fe551b7dba7d903f5dkccbq_495cf683-8b21-4b50-9207-b7727070c747/util/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.818704 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/util/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.979485 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/util/0.log" Nov 26 17:20:43 crc kubenswrapper[4956]: I1126 17:20:43.996686 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/pull/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.030762 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/pull/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.249586 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/util/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.269802 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/extract/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.291918 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_87b4bb7621dcb67338b53778f2871f07aa0e4d3dfcd0fd25724bfd240bn2cbw_11e5d606-14c5-459f-9549-f7ba6c5a698b/pull/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.452289 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/util/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.654926 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/pull/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.655120 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/pull/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.657323 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/util/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.834692 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/pull/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.840696 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/util/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.887854 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/util/0.log" Nov 26 17:20:44 crc kubenswrapper[4956]: I1126 17:20:44.916339 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9704761d240e56fb98655ffd81084895b33a73ec711f4dcdef0450e590zmm8x_73d12891-f813-4273-b9a8-de53a4067e14/extract/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.167492 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/util/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.183280 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/pull/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.193238 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/pull/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.367598 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/extract/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.371255 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/util/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.403954 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_9f0c59a3968beec894e04476dd5efd0a707bad85f482efd4940498368ccvblq_71a5370b-ba54-41ae-8e74-38a191f344d0/pull/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.421972 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/util/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.610520 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/util/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.612556 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/pull/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.635242 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/pull/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.799876 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/util/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.822793 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/extract/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.835521 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_d854280893f664a16f85f7c4268f877fa95509a4e25ae77fea242eaaa3f52m9_1c6c002c-d6a2-444f-bd42-c2ba47045304/pull/0.log" Nov 26 17:20:45 crc kubenswrapper[4956]: I1126 17:20:45.968092 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-788c9777f6-7qmf6_d2238f24-43e6-4512-880a-1c3b10ba3844/manager/3.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.026588 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-788c9777f6-7qmf6_d2238f24-43e6-4512-880a-1c3b10ba3844/manager/2.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.066174 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-index-9pjcm_e9e2f38e-3b35-4c00-ab0b-36602b28a4f0/registry-server/0.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.191064 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-dd7556475-2wzb2_5adb08fd-ab2a-4d2d-afab-4b605419b741/manager/3.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.220503 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-dd7556475-2wzb2_5adb08fd-ab2a-4d2d-afab-4b605419b741/manager/2.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.339695 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-index-kb7vl_f9b0e47e-e22e-4b2e-8b8c-732bb510c62d/registry-server/0.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.424252 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c4ffc67cb-dnf99_063794be-3564-44a8-8613-a966e3ebd953/kube-rbac-proxy/0.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.427980 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c4ffc67cb-dnf99_063794be-3564-44a8-8613-a966e3ebd953/manager/3.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.508291 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-5c4ffc67cb-dnf99_063794be-3564-44a8-8613-a966e3ebd953/manager/2.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.565060 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-index-t6ctx_24c5dfcb-c59d-4acf-b0c2-3d0f9d6f12af/registry-server/0.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.657040 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-69b487756b-xzvnf_7f723082-bb1f-4dd1-beb2-c1553728eba8/manager/3.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.721974 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-69b487756b-xzvnf_7f723082-bb1f-4dd1-beb2-c1553728eba8/manager/2.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.780104 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-index-phhw9_e70664dd-b470-4b7f-877c-8e7429a1ab7e/registry-server/0.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.893785 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7674ccc9cf-j5f9c_91276233-6cc3-408e-b9e4-e8a4c7ecc60d/manager/3.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.922081 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-7674ccc9cf-j5f9c_91276233-6cc3-408e-b9e4-e8a4c7ecc60d/manager/2.log" Nov 26 17:20:46 crc kubenswrapper[4956]: I1126 17:20:46.991853 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-index-w2wwv_cd46bd02-d445-4a8b-8712-b86b9ca6c0d2/registry-server/0.log" Nov 26 17:20:47 crc kubenswrapper[4956]: I1126 17:20:47.117145 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-94bjc_25d0fbbb-c310-46d0-ae91-a9804761e8a3/operator/3.log" Nov 26 17:20:47 crc kubenswrapper[4956]: I1126 17:20:47.143554 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-779fc9694b-94bjc_25d0fbbb-c310-46d0-ae91-a9804761e8a3/operator/2.log" Nov 26 17:20:47 crc kubenswrapper[4956]: I1126 17:20:47.197031 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-index-2nksr_fc6c140a-8045-4034-b13c-03972336f0de/registry-server/0.log" Nov 26 17:20:47 crc kubenswrapper[4956]: I1126 17:20:47.357155 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-f659d5cfc-2plpb_1090fc6e-22d8-4fb2-9296-46edde4e1af5/manager/2.log" Nov 26 17:20:47 crc kubenswrapper[4956]: I1126 17:20:47.367669 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-f659d5cfc-2plpb_1090fc6e-22d8-4fb2-9296-46edde4e1af5/manager/3.log" Nov 26 17:20:47 crc kubenswrapper[4956]: I1126 17:20:47.458739 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-index-rwrnv_434482ad-0b5b-47a5-aa36-289139e64252/registry-server/0.log" Nov 26 17:20:55 crc kubenswrapper[4956]: I1126 17:20:55.001747 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:20:55 crc kubenswrapper[4956]: E1126 17:20:55.002721 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:21:03 crc kubenswrapper[4956]: I1126 17:21:03.765987 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-gkbcz_4cf5e708-ff2f-4b3c-9653-1aace2d4bfff/control-plane-machine-set-operator/0.log" Nov 26 17:21:03 crc kubenswrapper[4956]: I1126 17:21:03.989214 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7vqnn_724bfa7f-010c-4298-a11a-50ea0f5fe198/kube-rbac-proxy/0.log" Nov 26 17:21:03 crc kubenswrapper[4956]: I1126 17:21:03.996259 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-7vqnn_724bfa7f-010c-4298-a11a-50ea0f5fe198/machine-api-operator/0.log" Nov 26 17:21:09 crc kubenswrapper[4956]: I1126 17:21:09.995035 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:21:09 crc kubenswrapper[4956]: E1126 17:21:09.996095 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:21:21 crc kubenswrapper[4956]: I1126 17:21:21.833302 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-rwzwq_1891d982-bc1a-4807-b53f-cf1837eb0c0b/kube-rbac-proxy/0.log" Nov 26 17:21:21 crc kubenswrapper[4956]: I1126 17:21:21.898485 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-6c7b4b5f48-rwzwq_1891d982-bc1a-4807-b53f-cf1837eb0c0b/controller/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.023076 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-frr-files/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.240659 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-frr-files/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.259221 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-reloader/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.275187 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-metrics/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.287601 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-reloader/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.451564 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-frr-files/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.475669 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-metrics/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.510138 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-reloader/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.524244 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-metrics/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.710160 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/controller/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.712975 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-reloader/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.723194 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-frr-files/0.log" Nov 26 17:21:22 crc kubenswrapper[4956]: I1126 17:21:22.770421 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/cp-metrics/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.074563 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/frr-metrics/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.074943 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/kube-rbac-proxy/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.136549 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/kube-rbac-proxy-frr/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.271565 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/reloader/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.494290 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-6998585d5-wnm7h_73efb92c-5732-443b-8f6f-a1b183f7a455/frr-k8s-webhook-server/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.599402 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d556748c5-l2bl4_4b5ac5a5-46f2-4fc2-944d-58fcafb997b5/manager/3.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.765029 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-79tsd_2ca64dff-1360-4261-81c8-61192f37b99a/frr/0.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.783157 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-6d556748c5-l2bl4_4b5ac5a5-46f2-4fc2-944d-58fcafb997b5/manager/2.log" Nov 26 17:21:23 crc kubenswrapper[4956]: I1126 17:21:23.814749 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-7bcbfbd867-h25xk_d7584f30-fba3-40ac-b6f2-881664bc6d6a/webhook-server/0.log" Nov 26 17:21:24 crc kubenswrapper[4956]: I1126 17:21:24.006453 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-768ll_86b7d34e-df87-4f72-a20b-d8f241516ac9/kube-rbac-proxy/0.log" Nov 26 17:21:24 crc kubenswrapper[4956]: I1126 17:21:24.230671 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-768ll_86b7d34e-df87-4f72-a20b-d8f241516ac9/speaker/0.log" Nov 26 17:21:25 crc kubenswrapper[4956]: I1126 17:21:25.002269 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:21:25 crc kubenswrapper[4956]: E1126 17:21:25.002672 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.324715 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-9e5d-account-create-update-qhfkw_ed6c8be2-8540-4937-8ee8-7209c72b1fbe/mariadb-account-create-update/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.512845 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-create-cz42g_6ee6b955-28a3-45e4-ba60-154ddd97ed1d/mariadb-database-create/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.589228 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-db-sync-8hkbb_eff77660-101d-4609-8ab9-21d93e94567f/glance-db-sync/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.695727 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-single-0_99d01d93-2c13-4500-ad93-e82c70e3df5d/glance-httpd/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.800423 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-single-0_99d01d93-2c13-4500-ad93-e82c70e3df5d/glance-log/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.825023 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-single-1_d6f63b3b-78e6-46cc-ab34-685c5deba6f4/glance-httpd/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.993456 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_glance-default-single-1_d6f63b3b-78e6-46cc-ab34-685c5deba6f4/glance-log/0.log" Nov 26 17:21:39 crc kubenswrapper[4956]: I1126 17:21:39.995409 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:21:39 crc kubenswrapper[4956]: E1126 17:21:39.995849 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:21:40 crc kubenswrapper[4956]: I1126 17:21:40.337342 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_13129b6d-631d-405b-8201-df79314c929c/mysql-bootstrap/0.log" Nov 26 17:21:40 crc kubenswrapper[4956]: I1126 17:21:40.491646 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_keystone-6c49f4fcbc-tchbn_6520d055-1a7e-47f7-a3c6-08d59d991dd8/keystone-api/0.log" Nov 26 17:21:40 crc kubenswrapper[4956]: I1126 17:21:40.585469 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_13129b6d-631d-405b-8201-df79314c929c/mysql-bootstrap/0.log" Nov 26 17:21:40 crc kubenswrapper[4956]: I1126 17:21:40.694335 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-0_13129b6d-631d-405b-8201-df79314c929c/galera/0.log" Nov 26 17:21:40 crc kubenswrapper[4956]: I1126 17:21:40.821604 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_6fc57249-9b8b-48d0-891b-36181a401d7a/mysql-bootstrap/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.020859 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_6fc57249-9b8b-48d0-891b-36181a401d7a/mysql-bootstrap/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.050317 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-1_6fc57249-9b8b-48d0-891b-36181a401d7a/galera/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.292005 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa/mysql-bootstrap/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.355894 4956 scope.go:117] "RemoveContainer" containerID="5fcb8669653662238df17c20f64712eae066c96abd77f821c064a0be03eed32a" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.383628 4956 scope.go:117] "RemoveContainer" containerID="5767b20b397a28d1c45dcf61e4de7c5a4a7f897fce042c2a43444107bb4e6e59" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.408237 4956 scope.go:117] "RemoveContainer" containerID="8b7ae5f26d68354fc34f1d985327638a2c9771cb3e36ad8276954a7d5f77a48d" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.446477 4956 scope.go:117] "RemoveContainer" containerID="18551c8bd2a6a38f12abf66c667eb04138078566eef08a87c1dc50ecb7afa55d" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.535122 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa/mysql-bootstrap/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.573569 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstack-galera-2_e7deb4ae-2c0e-41e3-9b94-b02426fdbaaa/galera/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.813180 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312/setup-container/0.log" Nov 26 17:21:41 crc kubenswrapper[4956]: I1126 17:21:41.862748 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_openstackclient_c193b63d-c72a-4f03-a669-2fd125152da6/openstackclient/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.074954 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312/rabbitmq/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.145840 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_rabbitmq-server-0_d4d157f9-7ae0-4d0c-97dd-75d2d7ecc312/setup-container/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.354654 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-bj6nd_f91290f1-6c5a-4199-806d-ab204e05814e/proxy-httpd/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.377120 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-proxy-6bd58cfcf7-bj6nd_f91290f1-6c5a-4199-806d-ab204e05814e/proxy-server/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.613152 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-ring-rebalance-r72t9_83a3f8ba-c224-4656-b7a6-8ac28c1804f9/swift-ring-rebalance/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.704317 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/account-auditor/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.888192 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/account-reaper/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.901733 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_memcached-0_199646a6-766c-4047-bced-5bada18d54bf/memcached/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.926158 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/account-replicator/0.log" Nov 26 17:21:42 crc kubenswrapper[4956]: I1126 17:21:42.942628 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/account-server/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.074570 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/container-auditor/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.133990 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/container-replicator/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.138165 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/container-server/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.197090 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/object-auditor/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.216350 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/container-updater/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.275690 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/object-expirer/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.321683 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/object-replicator/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.353947 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/object-server/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.422500 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/object-updater/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.430363 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/rsync/0.log" Nov 26 17:21:43 crc kubenswrapper[4956]: I1126 17:21:43.483524 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/glance-kuttl-tests_swift-storage-0_26554e9d-3393-425b-bc71-07e66397ca3b/swift-recon-cron/0.log" Nov 26 17:21:53 crc kubenswrapper[4956]: I1126 17:21:53.995793 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:21:53 crc kubenswrapper[4956]: E1126 17:21:53.996789 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:21:57 crc kubenswrapper[4956]: I1126 17:21:57.462283 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/extract-utilities/0.log" Nov 26 17:21:57 crc kubenswrapper[4956]: I1126 17:21:57.593933 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/extract-content/0.log" Nov 26 17:21:57 crc kubenswrapper[4956]: I1126 17:21:57.600185 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/extract-utilities/0.log" Nov 26 17:21:57 crc kubenswrapper[4956]: I1126 17:21:57.600671 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/extract-content/0.log" Nov 26 17:21:57 crc kubenswrapper[4956]: I1126 17:21:57.921802 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/extract-content/0.log" Nov 26 17:21:57 crc kubenswrapper[4956]: I1126 17:21:57.922013 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/extract-utilities/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.104748 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/extract-utilities/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.420036 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/extract-content/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.430089 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/extract-content/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.450451 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/extract-utilities/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.520137 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-kspnz_520be241-f85c-43e0-8fe0-892e0df6c66c/registry-server/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.607933 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/extract-utilities/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.648127 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/extract-content/0.log" Nov 26 17:21:58 crc kubenswrapper[4956]: I1126 17:21:58.855469 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/util/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.038764 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-vgxbf_5102435b-0e33-4be2-a344-df54bd02afa6/registry-server/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.055225 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/util/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.118602 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/pull/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.119980 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/pull/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.297740 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/util/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.319503 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/pull/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.341482 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_e8527aae5664f20f24bf3bbb3fd2981ba838928a8a47ce599ee258e4c658jlv_5da5ddac-6e96-4800-b178-a03b3cdc1dfa/extract/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.484105 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-8drrp_c33f19c2-c49b-4677-805f-e60441939c00/marketplace-operator/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.513969 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/extract-utilities/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.709683 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/extract-content/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.709689 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/extract-content/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.717929 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/extract-utilities/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.918574 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/extract-utilities/0.log" Nov 26 17:21:59 crc kubenswrapper[4956]: I1126 17:21:59.934729 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/extract-content/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.007945 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-c6bhv_abad77b8-1ee5-455d-9510-f76fbc9f3014/registry-server/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.138919 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/extract-utilities/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.314306 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/extract-utilities/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.316552 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/extract-content/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.338945 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/extract-content/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.509675 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/extract-content/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.521008 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/extract-utilities/0.log" Nov 26 17:22:00 crc kubenswrapper[4956]: I1126 17:22:00.947035 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-mvl9d_ab83ee79-8dfe-4cd9-b104-49ac79683da9/registry-server/0.log" Nov 26 17:22:05 crc kubenswrapper[4956]: I1126 17:22:05.000619 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:22:05 crc kubenswrapper[4956]: E1126 17:22:05.001470 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:22:15 crc kubenswrapper[4956]: I1126 17:22:15.995952 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:22:16 crc kubenswrapper[4956]: E1126 17:22:15.997056 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:22:29 crc kubenswrapper[4956]: I1126 17:22:29.995125 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:22:29 crc kubenswrapper[4956]: E1126 17:22:29.996110 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:22:41 crc kubenswrapper[4956]: I1126 17:22:41.519302 4956 scope.go:117] "RemoveContainer" containerID="a18e29b90b2f5c83e72cf855dd4f468da0b9e982cd4ed16c6a74511f7fa3e9a6" Nov 26 17:22:41 crc kubenswrapper[4956]: I1126 17:22:41.591328 4956 scope.go:117] "RemoveContainer" containerID="46b2afe6fccde0256021d6e3ddf58a8ed8a149d4b75752394c111b7e15c674f5" Nov 26 17:22:41 crc kubenswrapper[4956]: I1126 17:22:41.628057 4956 scope.go:117] "RemoveContainer" containerID="ef8f20a63088a1211acf6f328879a9ee13c75df423b8d7b70dbe43264d39713b" Nov 26 17:22:41 crc kubenswrapper[4956]: I1126 17:22:41.672660 4956 scope.go:117] "RemoveContainer" containerID="5e56230a8c816380d0c0f85d11a40729b126bb9da2304acba577e7689001cf70" Nov 26 17:22:45 crc kubenswrapper[4956]: I1126 17:22:45.005686 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:22:45 crc kubenswrapper[4956]: E1126 17:22:45.007027 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:22:56 crc kubenswrapper[4956]: I1126 17:22:56.996233 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:22:56 crc kubenswrapper[4956]: E1126 17:22:56.997301 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:23:09 crc kubenswrapper[4956]: I1126 17:23:09.995474 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:23:09 crc kubenswrapper[4956]: E1126 17:23:09.996585 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:23:13 crc kubenswrapper[4956]: I1126 17:23:13.335132 4956 generic.go:334] "Generic (PLEG): container finished" podID="567fade8-9828-483f-b993-261298068736" containerID="b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9" exitCode=0 Nov 26 17:23:13 crc kubenswrapper[4956]: I1126 17:23:13.335244 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-9l4hp/must-gather-92ljh" event={"ID":"567fade8-9828-483f-b993-261298068736","Type":"ContainerDied","Data":"b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9"} Nov 26 17:23:13 crc kubenswrapper[4956]: I1126 17:23:13.336159 4956 scope.go:117] "RemoveContainer" containerID="b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9" Nov 26 17:23:14 crc kubenswrapper[4956]: I1126 17:23:14.343056 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9l4hp_must-gather-92ljh_567fade8-9828-483f-b993-261298068736/gather/0.log" Nov 26 17:23:21 crc kubenswrapper[4956]: I1126 17:23:21.914160 4956 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-9l4hp/must-gather-92ljh"] Nov 26 17:23:21 crc kubenswrapper[4956]: I1126 17:23:21.915874 4956 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-9l4hp/must-gather-92ljh" podUID="567fade8-9828-483f-b993-261298068736" containerName="copy" containerID="cri-o://5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa" gracePeriod=2 Nov 26 17:23:21 crc kubenswrapper[4956]: I1126 17:23:21.930502 4956 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-9l4hp/must-gather-92ljh"] Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.319312 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9l4hp_must-gather-92ljh_567fade8-9828-483f-b993-261298068736/copy/0.log" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.320192 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.344244 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/567fade8-9828-483f-b993-261298068736-must-gather-output\") pod \"567fade8-9828-483f-b993-261298068736\" (UID: \"567fade8-9828-483f-b993-261298068736\") " Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.344334 4956 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-79mq7\" (UniqueName: \"kubernetes.io/projected/567fade8-9828-483f-b993-261298068736-kube-api-access-79mq7\") pod \"567fade8-9828-483f-b993-261298068736\" (UID: \"567fade8-9828-483f-b993-261298068736\") " Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.366808 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/567fade8-9828-483f-b993-261298068736-kube-api-access-79mq7" (OuterVolumeSpecName: "kube-api-access-79mq7") pod "567fade8-9828-483f-b993-261298068736" (UID: "567fade8-9828-483f-b993-261298068736"). InnerVolumeSpecName "kube-api-access-79mq7". PluginName "kubernetes.io/projected", VolumeGidValue "" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.424809 4956 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-9l4hp_must-gather-92ljh_567fade8-9828-483f-b993-261298068736/copy/0.log" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.425435 4956 generic.go:334] "Generic (PLEG): container finished" podID="567fade8-9828-483f-b993-261298068736" containerID="5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa" exitCode=143 Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.425519 4956 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-9l4hp/must-gather-92ljh" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.425539 4956 scope.go:117] "RemoveContainer" containerID="5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.440219 4956 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/567fade8-9828-483f-b993-261298068736-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "567fade8-9828-483f-b993-261298068736" (UID: "567fade8-9828-483f-b993-261298068736"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.446152 4956 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/567fade8-9828-483f-b993-261298068736-must-gather-output\") on node \"crc\" DevicePath \"\"" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.446186 4956 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-79mq7\" (UniqueName: \"kubernetes.io/projected/567fade8-9828-483f-b993-261298068736-kube-api-access-79mq7\") on node \"crc\" DevicePath \"\"" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.453440 4956 scope.go:117] "RemoveContainer" containerID="b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.506975 4956 scope.go:117] "RemoveContainer" containerID="5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa" Nov 26 17:23:22 crc kubenswrapper[4956]: E1126 17:23:22.507730 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa\": container with ID starting with 5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa not found: ID does not exist" containerID="5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.507785 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa"} err="failed to get container status \"5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa\": rpc error: code = NotFound desc = could not find container \"5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa\": container with ID starting with 5a63b9402b84f5f1e47c984443e45ccca141d4360e925373745fd1afc43282aa not found: ID does not exist" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.507822 4956 scope.go:117] "RemoveContainer" containerID="b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9" Nov 26 17:23:22 crc kubenswrapper[4956]: E1126 17:23:22.508484 4956 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9\": container with ID starting with b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9 not found: ID does not exist" containerID="b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9" Nov 26 17:23:22 crc kubenswrapper[4956]: I1126 17:23:22.508536 4956 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9"} err="failed to get container status \"b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9\": rpc error: code = NotFound desc = could not find container \"b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9\": container with ID starting with b5130088effa9191297b196839525aa056a9bf9c1f8ab48a36d1baa373dc98d9 not found: ID does not exist" Nov 26 17:23:23 crc kubenswrapper[4956]: I1126 17:23:23.004659 4956 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="567fade8-9828-483f-b993-261298068736" path="/var/lib/kubelet/pods/567fade8-9828-483f-b993-261298068736/volumes" Nov 26 17:23:23 crc kubenswrapper[4956]: I1126 17:23:23.995630 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:23:23 crc kubenswrapper[4956]: E1126 17:23:23.996043 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:23:37 crc kubenswrapper[4956]: I1126 17:23:37.996434 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:23:37 crc kubenswrapper[4956]: E1126 17:23:37.997880 4956 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-9rvrb_openshift-machine-config-operator(fadaf3cf-cfc5-4f60-bd4a-4eae814da018)\"" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" Nov 26 17:23:51 crc kubenswrapper[4956]: I1126 17:23:51.995504 4956 scope.go:117] "RemoveContainer" containerID="3aeff1ae3ff87e70f20b38f341ecdcd2f214649527ba3595b10e0d94fd9e0186" Nov 26 17:23:52 crc kubenswrapper[4956]: I1126 17:23:52.699289 4956 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" event={"ID":"fadaf3cf-cfc5-4f60-bd4a-4eae814da018","Type":"ContainerStarted","Data":"6e01d6ca6a1dd5feec624b9977b100ef70b11711dcd4781f167d723fb9195ebb"} Nov 26 17:24:41 crc kubenswrapper[4956]: I1126 17:24:41.829196 4956 scope.go:117] "RemoveContainer" containerID="1266462912a885656e1f047d08ab65c6e573ae1f47cc909452e5a33df74a577c" Nov 26 17:26:09 crc kubenswrapper[4956]: I1126 17:26:09.551015 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:26:09 crc kubenswrapper[4956]: I1126 17:26:09.551733 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Nov 26 17:26:39 crc kubenswrapper[4956]: I1126 17:26:39.551803 4956 patch_prober.go:28] interesting pod/machine-config-daemon-9rvrb container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Nov 26 17:26:39 crc kubenswrapper[4956]: I1126 17:26:39.552678 4956 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-9rvrb" podUID="fadaf3cf-cfc5-4f60-bd4a-4eae814da018" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515111634157024451 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015111634160017360 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015111627670016513 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015111627671015464 5ustar corecore